Anforderungen  |   Konzepte  |   Entwurf  |   Entwicklung  |   Qualitätssicherung  |   Lebenszyklus  |   Steuerung
 
 
 
 


Quelle  run_vmtests.sh   Sprache: Shell

 
#!/bin/bash
# SPDX-License-Identifier: GPL-2.0
# Please run as root

# Kselftest framework requirement - SKIP code is 4.
ksft_skip=4

count_total=0
count_pass=0
count_fail=0
count_skip=0
exitcode=0

usage() {
 cat <<EOF
usage: ${BASH_SOURCE[0]:-$0} [ options ]

  -a: run all tests, including extra ones (other than destructive ones)
  -t: specify specific categories to tests to run
  -h: display this message
  -n: disable TAP output
  -d: run destructive tests

The default behavior is to run required tests only.  If -a is specified,
will run all tests.

Alternatively, specific groups tests can be run by passing a string
to the -t argument containing one or more of the following categories
separated by spaces:
- mmap
 tests for mmap(2)
- gup_test
 tests for gup
- userfaultfd
 tests for  userfaultfd(2)
- compaction
 a test for the patch "Allow compaction of unevictable pages"
- mlock
 tests for mlock(2)
- mremap
 tests for mremap(2)
- hugevm
 tests for very large virtual address space
- vmalloc
 vmalloc smoke tests
- hmm
 hmm smoke tests
- madv_guard
 test madvise(2) MADV_GUARD_INSTALL and MADV_GUARD_REMOVE options
- madv_populate
 test memadvise(2) MADV_POPULATE_{READ,WRITE} options
- memfd_secret
 test memfd_secret(2)
- process_mrelease
 test process_mrelease(2)
- ksm
 ksm tests that do not require >=2 NUMA nodes
- ksm_numa
 ksm tests that require >=2 NUMA nodes
- pkey
 memory protection key tests
- soft_dirty
 test soft dirty page bit semantics
- pagemap
 test pagemap_scan IOCTL
- pfnmap
 tests for VM_PFNMAP handling
- process_madv
 test for process_madv
- cow
 test copy-on-write semantics
- thp
 test transparent huge pages
- hugetlb
 test hugetlbfs huge pages
- migration
 invoke move_pages(2) to exercise the migration entry code
 paths in the kernel
- mkdirty
 test handling of code that might set PTE/PMD dirty in
 read-only VMAs
- mdwe
 test prctl(PR_SET_MDWE, ...)
- page_frag
 test handling of page fragment allocation and freeing
- vma_merge
 test VMA merge cases behave as expected

example: ./run_vmtests.sh -t "hmm mmap ksm"
EOF
 exit 0
}

RUN_ALL=false
RUN_DESTRUCTIVE=false
TAP_PREFIX="# "

while getopts "aht:n" OPT; do
 case ${OPT} in
  "a") RUN_ALL=true ;;
  "h") usage ;;
  "t") VM_SELFTEST_ITEMS=${OPTARG} ;;
  "n") TAP_PREFIX= ;;
  "d") RUN_DESTRUCTIVE=true ;;
 esac
done
shift $((OPTIND -1))

# default behavior: run all tests
VM_SELFTEST_ITEMS=${VM_SELFTEST_ITEMS:-default}

test_selected() {
 if [ "$VM_SELFTEST_ITEMS" == "default" ]; then
  # If no VM_SELFTEST_ITEMS are specified, run all tests
  return 0
 fi
 # If test selected argument is one of the test items
 if [[ " ${VM_SELFTEST_ITEMS[*]} " =~ " ${1} " ]]; then
         return 0
 else
         return 1
 fi
}

run_gup_matrix() {
    # -t: thp=on, -T: thp=off, -H: hugetlb=on
    local hugetlb_mb=$(( needmem_KB / 1024 ))

    for huge in -t -T "-H -m $hugetlb_mb"do
        # -u: gup-fast, -U: gup-basic, -a: pin-fast, -b: pin-basic, -L: pin-longterm
        for test_cmd in -u -U -a -b -L; do
            # -w: write=1, -W: write=0
            for write in -w -W; do
                # -S: shared
                for share in -S " "do
                    # -n: How many pages to fetch together?  512 is special
                    # because it's default thp size (or 2M on x86), 123 to
                    # just test partial gup when hit a huge in whatever form
                    for num in "-n 1" "-n 512" "-n 123"do
                        CATEGORY="gup_test" run_test ./gup_test \
                                $huge $test_cmd $write $share $num
                    done
                done
            done
        done
    done
}

# get huge pagesize and freepages from /proc/meminfo
while read -r name size unit; do
 if [ "$name" = "HugePages_Free:" ]; then
  freepgs="$size"
 fi
 if [ "$name" = "Hugepagesize:" ]; then
  hpgsize_KB="$size"
 fi
done < /proc/meminfo

# Simple hugetlbfs tests have a hardcoded minimum requirement of
# huge pages totaling 256MB (262144KB) in size.  The userfaultfd
# hugetlb test requires a minimum of 2 * nr_cpus huge pages.  Take
# both of these requirements into account and attempt to increase
# number of huge pages available.
nr_cpus=$(nproc)
uffd_min_KB=$((hpgsize_KB * nr_cpus * 2))
hugetlb_min_KB=$((256 * 1024))
if [[ $uffd_min_KB -gt $hugetlb_min_KB ]]; then
 needmem_KB=$uffd_min_KB
else
 needmem_KB=$hugetlb_min_KB
fi

# set proper nr_hugepages
if [ -n "$freepgs" ] && [ -n "$hpgsize_KB" ]; then
 nr_hugepgs=$(cat /proc/sys/vm/nr_hugepages)
 needpgs=$((needmem_KB / hpgsize_KB))
 tries=2
 while [ "$tries" -gt 0 ] && [ "$freepgs" -lt "$needpgs" ]; do
  lackpgs=$((needpgs - freepgs))
  echo 3 > /proc/sys/vm/drop_caches
  if ! echo $((lackpgs + nr_hugepgs)) > /proc/sys/vm/nr_hugepages; then
   echo "Please run this test as root"
   exit $ksft_skip
  fi
  while read -r name size unit; do
   if [ "$name" = "HugePages_Free:" ]; then
    freepgs=$size
   fi
  done < /proc/meminfo
  tries=$((tries - 1))
 done
 if [ "$freepgs" -lt "$needpgs" ]; then
  printf "Not enough huge pages available (%d < %d)\n" \
         "$freepgs" "$needpgs"
 fi
 HAVE_HUGEPAGES=1
else
 echo "no hugetlbfs support in kernel?"
 HAVE_HUGEPAGES=0
fi

# filter 64bit architectures
ARCH64STR="arm64 mips64 parisc64 ppc64 ppc64le riscv64 s390x sparc64 x86_64"
if [ -z "$ARCH" ]; then
 ARCH=$(uname -m 2>/dev/null | sed -e 's/aarch64.*/arm64/')
fi
VADDR64=0
echo "$ARCH64STR" | grep "$ARCH" &>/dev/null && VADDR64=1

tap_prefix() {
 sed -e "s/^/${TAP_PREFIX}/"
}

tap_output() {
 if [[ ! -z "$TAP_PREFIX" ]]; then
  read str
  echo $str
 fi
}

pretty_name() {
 echo "$*" | sed -e 's/^\(bash \)\?\.\///'
}

# Usage: run_test [test binary] [arbitrary test arguments...]
run_test() {
 if test_selected ${CATEGORY}; then
  local skip=0

  # On memory constrainted systems some tests can fail to allocate hugepages.
  # perform some cleanup before the test for a higher success rate.
  if [ ${CATEGORY} == "thp" -o ${CATEGORY} == "hugetlb" ]; then
   if [ "${HAVE_HUGEPAGES}" = "1" ]; then
    echo 3 > /proc/sys/vm/drop_caches
    sleep 2
    echo 1 > /proc/sys/vm/compact_memory
    sleep 2
   else
    echo "hugepages not supported" | tap_prefix
    skip=1
   fi
  fi

  local test=$(pretty_name "$*")
  local title="running $*"
  local sep=$(echo -n "$title" | tr "[:graph:][:space:]" -)
  printf "%s\n%s\n%s\n" "$sep" "$title" "$sep" | tap_prefix

  if [ "${skip}" != "1" ]; then
   ("$@" 2>&1) | tap_prefix
   local ret=${PIPESTATUS[0]}
  else
   local ret=$ksft_skip
  fi
  count_total=$(( count_total + 1 ))
  if [ $ret -eq 0 ]; then
   count_pass=$(( count_pass + 1 ))
   echo "[PASS]" | tap_prefix
   echo "ok ${count_total} ${test}" | tap_output
  elif [ $ret -eq $ksft_skip ]; then
   count_skip=$(( count_skip + 1 ))
   echo "[SKIP]" | tap_prefix
   echo "ok ${count_total} ${test} # SKIP" | tap_output
   exitcode=$ksft_skip
  else
   count_fail=$(( count_fail + 1 ))
   echo "[FAIL]" | tap_prefix
   echo "not ok ${count_total} ${test} # exit=$ret" | tap_output
   exitcode=1
  fi
 fi # test_selected
}

echo "TAP version 13" | tap_output

CATEGORY="hugetlb" run_test ./hugepage-mmap

shmmax=$(cat /proc/sys/kernel/shmmax)
shmall=$(cat /proc/sys/kernel/shmall)
echo 268435456 > /proc/sys/kernel/shmmax
echo 4194304 > /proc/sys/kernel/shmall
CATEGORY="hugetlb" run_test ./hugepage-shm
echo "$shmmax" > /proc/sys/kernel/shmmax
echo "$shmall" > /proc/sys/kernel/shmall

CATEGORY="hugetlb" run_test ./map_hugetlb
CATEGORY="hugetlb" run_test ./hugepage-mremap
CATEGORY="hugetlb" run_test ./hugepage-vmemmap
CATEGORY="hugetlb" run_test ./hugetlb-madvise
CATEGORY="hugetlb" run_test ./hugetlb_dio

if [ "${HAVE_HUGEPAGES}" = "1" ]; then
 nr_hugepages_tmp=$(cat /proc/sys/vm/nr_hugepages)
 # For this test, we need one and just one huge page
 echo 1 > /proc/sys/vm/nr_hugepages
 CATEGORY="hugetlb" run_test ./hugetlb_fault_after_madv
 CATEGORY="hugetlb" run_test ./hugetlb_madv_vs_map
 # Restore the previous number of huge pages, since further tests rely on it
 echo "$nr_hugepages_tmp" > /proc/sys/vm/nr_hugepages
fi

if test_selected "hugetlb"then
 echo "NOTE: These hugetlb tests provide minimal coverage. Use"   | tap_prefix
 echo " https://github.com/libhugetlbfs/libhugetlbfs.git for" | tap_prefix
 echo " hugetlb regression testing."     | tap_prefix
fi

CATEGORY="mmap" run_test ./map_fixed_noreplace

if $RUN_ALL; then
    run_gup_matrix
else
    # get_user_pages_fast() benchmark
    CATEGORY="gup_test" run_test ./gup_test -u
    # pin_user_pages_fast() benchmark
    CATEGORY="gup_test" run_test ./gup_test -a
fi
# Dump pages 0, 19, and 4096, using pin_user_pages:
CATEGORY="gup_test" run_test ./gup_test -ct -F 0x1 0 19 0x1000
CATEGORY="gup_test" run_test ./gup_longterm

CATEGORY="userfaultfd" run_test ./uffd-unit-tests
uffd_stress_bin=./uffd-stress
CATEGORY="userfaultfd" run_test ${uffd_stress_bin} anon 20 16
# Hugetlb tests require source and destination huge pages. Pass in half
# the size of the free pages we have, which is used for *each*.
# uffd-stress expects a region expressed in MiB, so we adjust
# half_ufd_size_MB accordingly.
half_ufd_size_MB=$(((freepgs * hpgsize_KB) / 1024 / 2))
CATEGORY="userfaultfd" run_test ${uffd_stress_bin} hugetlb "$half_ufd_size_MB" 32
CATEGORY="userfaultfd" run_test ${uffd_stress_bin} hugetlb-private "$half_ufd_size_MB" 32
CATEGORY="userfaultfd" run_test ${uffd_stress_bin} shmem 20 16
CATEGORY="userfaultfd" run_test ${uffd_stress_bin} shmem-private 20 16
# uffd-wp-mremap requires at least one page of each size.
have_all_size_hugepgs=true
declare -A nr_size_hugepgs
for f in /sys/kernel/mm/hugepages/**/nr_hugepages; do
 old=$(cat $f)
 nr_size_hugepgs["$f"]="$old"
 if [ "$old" == 0 ]; then
  echo 1 > "$f"
 fi
 if [ $(cat "$f") == 0 ]; then
  have_all_size_hugepgs=false
  break
 fi
done
if $have_all_size_hugepgs; then
 CATEGORY="userfaultfd" run_test ./uffd-wp-mremap
else
 echo "# SKIP ./uffd-wp-mremap"
fi

#cleanup
for f in "${!nr_size_hugepgs[@]}"do
 echo "${nr_size_hugepgs["$f"]}" > "$f"
done
echo "$nr_hugepgs" > /proc/sys/vm/nr_hugepages

CATEGORY="compaction" run_test ./compaction_test

if command -v sudo &> /dev/null && sudo -u nobody ls ./on-fault-limit >/dev/null;
then
 CATEGORY="mlock" run_test sudo -u nobody ./on-fault-limit
else
 echo "# SKIP ./on-fault-limit"
fi

CATEGORY="mmap" run_test ./map_populate

CATEGORY="mlock" run_test ./mlock-random-test

CATEGORY="mlock" run_test ./mlock2-tests

CATEGORY="process_mrelease" run_test ./mrelease_test

CATEGORY="mremap" run_test ./mremap_test

CATEGORY="hugetlb" run_test ./thuge-gen
CATEGORY="hugetlb" run_test ./charge_reserved_hugetlb.sh -cgroup-v2
CATEGORY="hugetlb" run_test ./hugetlb_reparenting_test.sh -cgroup-v2
if $RUN_DESTRUCTIVE; then
nr_hugepages_tmp=$(cat /proc/sys/vm/nr_hugepages)
enable_soft_offline=$(cat /proc/sys/vm/enable_soft_offline)
echo 8 > /proc/sys/vm/nr_hugepages
CATEGORY="hugetlb" run_test ./hugetlb-soft-offline
echo "$nr_hugepages_tmp" > /proc/sys/vm/nr_hugepages
echo "$enable_soft_offline" > /proc/sys/vm/enable_soft_offline
CATEGORY="hugetlb" run_test ./hugetlb-read-hwpoison
fi

if [ $VADDR64 -ne 0 ]; then

 # set overcommit_policy as OVERCOMMIT_ALWAYS so that kernel
 # allows high virtual address allocation requests independent
 # of platform's physical memory.

 if [ -x ./virtual_address_range ]; then
  prev_policy=$(cat /proc/sys/vm/overcommit_memory)
  echo 1 > /proc/sys/vm/overcommit_memory
  CATEGORY="hugevm" run_test ./virtual_address_range
  echo $prev_policy > /proc/sys/vm/overcommit_memory
 fi

 # va high address boundary switch test
 ARCH_ARM64="arm64"
 prev_nr_hugepages=$(cat /proc/sys/vm/nr_hugepages)
 if [ "$ARCH" == "$ARCH_ARM64" ]; then
  echo 6 > /proc/sys/vm/nr_hugepages
 fi
 CATEGORY="hugevm" run_test bash ./va_high_addr_switch.sh
 if [ "$ARCH" == "$ARCH_ARM64" ]; then
  echo $prev_nr_hugepages > /proc/sys/vm/nr_hugepages
 fi
fi # VADDR64

# vmalloc stability smoke test
CATEGORY="vmalloc" run_test bash ./test_vmalloc.sh smoke

CATEGORY="mremap" run_test ./mremap_dontunmap

CATEGORY="hmm" run_test bash ./test_hmm.sh smoke

# MADV_GUARD_INSTALL and MADV_GUARD_REMOVE tests
CATEGORY="madv_guard" run_test ./guard-regions

# MADV_POPULATE_READ and MADV_POPULATE_WRITE tests
CATEGORY="madv_populate" run_test ./madv_populate

# PROCESS_MADV test
CATEGORY="process_madv" run_test ./process_madv

CATEGORY="vma_merge" run_test ./merge

if [ -x ./memfd_secret ]
then
if [ -f /proc/sys/kernel/yama/ptrace_scope ]; then
 (echo 0 > /proc/sys/kernel/yama/ptrace_scope 2>&1) | tap_prefix
fi
CATEGORY="memfd_secret" run_test ./memfd_secret
fi

# KSM KSM_MERGE_TIME_HUGE_PAGES test with size of 100
if [ "${HAVE_HUGEPAGES}" = "1" ]; then
 CATEGORY="ksm" run_test ./ksm_tests -H -s 100
fi
# KSM KSM_MERGE_TIME test with size of 100
CATEGORY="ksm" run_test ./ksm_tests -P -s 100
# KSM MADV_MERGEABLE test with 10 identical pages
CATEGORY="ksm" run_test ./ksm_tests -M -p 10
# KSM unmerge test
CATEGORY="ksm" run_test ./ksm_tests -U
# KSM test with 10 zero pages and use_zero_pages = 0
CATEGORY="ksm" run_test ./ksm_tests -Z -p 10 -z 0
# KSM test with 10 zero pages and use_zero_pages = 1
CATEGORY="ksm" run_test ./ksm_tests -Z -p 10 -z 1
# KSM test with 2 NUMA nodes and merge_across_nodes = 1
CATEGORY="ksm_numa" run_test ./ksm_tests -N -m 1
# KSM test with 2 NUMA nodes and merge_across_nodes = 0
CATEGORY="ksm_numa" run_test ./ksm_tests -N -m 0

CATEGORY="ksm" run_test ./ksm_functional_tests

# protection_keys tests
nr_hugepgs=$(cat /proc/sys/vm/nr_hugepages)
if [ -x ./protection_keys_32 ]
then
 CATEGORY="pkey" run_test ./protection_keys_32
fi

if [ -x ./protection_keys_64 ]
then
 CATEGORY="pkey" run_test ./protection_keys_64
fi
echo "$nr_hugepgs" > /proc/sys/vm/nr_hugepages

if [ -x ./soft-dirty ]
then
 CATEGORY="soft_dirty" run_test ./soft-dirty
fi

CATEGORY="pagemap" run_test ./pagemap_ioctl

CATEGORY="pfnmap" run_test ./pfnmap

# COW tests
CATEGORY="cow" run_test ./cow

CATEGORY="thp" run_test ./khugepaged

CATEGORY="thp" run_test ./khugepaged -s 2

CATEGORY="thp" run_test ./khugepaged all:shmem

CATEGORY="thp" run_test ./khugepaged -s 4 all:shmem

CATEGORY="thp" run_test ./transhuge-stress -d 20

# Try to create XFS if not provided
if [ -z "${SPLIT_HUGE_PAGE_TEST_XFS_PATH}" ]; then
    if [ "${HAVE_HUGEPAGES}" = "1" ]; then
 if test_selected "thp"then
     if grep xfs /proc/filesystems &>/dev/null; then
  XFS_IMG=$(mktemp /tmp/xfs_img_XXXXXX)
  SPLIT_HUGE_PAGE_TEST_XFS_PATH=$(mktemp -d /tmp/xfs_dir_XXXXXX)
  truncate -s 314572800 ${XFS_IMG}
  mkfs.xfs -q ${XFS_IMG}
  mount -o loop ${XFS_IMG} ${SPLIT_HUGE_PAGE_TEST_XFS_PATH}
  MOUNTED_XFS=1
     fi
 fi
    fi
fi

CATEGORY="thp" run_test ./split_huge_page_test ${SPLIT_HUGE_PAGE_TEST_XFS_PATH}

if [ -n "${MOUNTED_XFS}" ]; then
    umount ${SPLIT_HUGE_PAGE_TEST_XFS_PATH}
    rmdir ${SPLIT_HUGE_PAGE_TEST_XFS_PATH}
    rm -f ${XFS_IMG}
fi

CATEGORY="migration" run_test ./migration

CATEGORY="mkdirty" run_test ./mkdirty

CATEGORY="mdwe" run_test ./mdwe_test

CATEGORY="page_frag" run_test ./test_page_frag.sh smoke

CATEGORY="page_frag" run_test ./test_page_frag.sh aligned

CATEGORY="page_frag" run_test ./test_page_frag.sh nonaligned

echo "SUMMARY: PASS=${count_pass} SKIP=${count_skip} FAIL=${count_fail}" | tap_prefix
echo "1..${count_total}" | tap_output

exit $exitcode

Messung V0.5
C=93 H=90 G=91

¤ Dauer der Verarbeitung: 0.25 Sekunden  (vorverarbeitet)  ¤

*© Formatika GbR, Deutschland






Wurzel

Suchen

Beweissystem der NASA

Beweissystem Isabelle

NIST Cobol Testsuite

Cephes Mathematical Library

Wiener Entwicklungsmethode

Haftungshinweis

Die Informationen auf dieser Webseite wurden nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit, noch Qualität der bereit gestellten Informationen zugesichert.

Bemerkung:

Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.






                                                                                                                                                                                                                                                                                                                                                                                                     


Neuigkeiten

     Aktuelles
     Motto des Tages

Software

     Produkte
     Quellcodebibliothek

Aktivitäten

     Artikel über Sicherheit
     Anleitung zur Aktivierung von SSL

Muße

     Gedichte
     Musik
     Bilder

Jenseits des Üblichen ....
    

Besucherstatistik

Besucherstatistik

Monitoring

Montastic status badge