Found 25 repositories(showing 25)
CNTRUN
char const* const commands[] = { "aapt", " aapt", " zipalign", "abduco", " abduco", "abook", " abook", "alpine", " alpine", " pico", " pilot", " rpdump", " rpload", "angband", " angband", "apache2", " ab", " apachectl", " apxs", " checkgid", " dbmmanage", " envvars-std", " fcgistarter", " htcacheclean", " htdbm", " htdigest", " htpasswd", " httpd", " httxt2dbm", " logresolve", " rotatelogs", " suexec", "apr-dev", " apr-1-config", "apr-util-dev", " apu-1-config", "apt", " apt", " apt-cache", " apt-config", " apt-get", " apt-key", " apt-mark", "aria2", " aria2c", "atomicparsley", " AtomicParsley", "attr", " attr", " getfattr", " setfattr", "autossh", " autossh", "bash", " bash", "bc", " bc", " dc", "binutils", " addr2line", " ar", " arm-linux-androideabi-ar", " arm-linux-androideabi-ld", " arm-linux-androideabi-nm", " arm-linux-androideabi-objdump", " arm-linux-androideabi-ranlib", " arm-linux-androideabi-readelf", " arm-linux-androideabi-strip", " as", " c++filt", " elfedit", " gprof", " ld", " ldd", " nm", " objcopy", " objdump", " ranlib", " readelf", " size", " strings", " strip", "bison", " bison", " yacc", "blogc", " blogc", " blogc-make", " blogc-runserver", "bmon", " bmon", "brogue", " brogue", "bs1770gain", " bs1770gain", "bsdtar", " bsdcat", " bsdcpio", " bsdtar", "busybox", " busybox", " env", "bvi", " bmore", " bvedit", " bvi", " bview", "bzip2", " bunzip2", " bzcat", " bzcmp", " bzdiff", " bzgrep", " bzip2", " bzip2recover", " bzless", " bzmore", "cadaver", " cadaver", "calcurse", " calcurse", " calcurse-caldav", " calcurse-upgrade", "cava", " cava", "cboard", " cboard", "ccache", " ccache", "ccrypt", " ccat", " ccdecrypt", " ccencrypt", " ccguess", " ccrypt", "cgdb", " cgdb", "clang", " arm-linux-androideabi-clang", " arm-linux-androideabi-clang++", " arm-linux-androideabi-cpp", " arm-linux-androideabi-g++", " arm-linux-androideabi-gcc", " c++", " cc", " clang", " clang++", " clang-5.0", " clang-cl", " clang-cpp", " clang-format", " clang-rename", " cpp", " g++", " gcc", "cmake", " cmake", " cpack", " ctest", "cmake-curses-gui", " ccmake", "cmark", " cmark", "cmatrix", " cmatrix", "cmus", " cmus", " cmus-remote", "coreutils", " [", " b2sum", " base32", " base64", " basename", " cat", " chcon", " chgrp", " chmod", " chown", " cksum", " comm", " coreutils", " cp", " csplit", " cut", " date", " dd", " dir", " dircolors", " dirname", " du", " echo", " expand", " expr", " factor", " false", " fmt", " fold", " groups", " head", " id", " install", " join", " kill", " link", " ln", " logname", " ls", " md5sum", " mkdir", " mkfifo", " mknod", " mktemp", " mv", " nice", " nl", " nohup", " nproc", " numfmt", " od", " paste", " pathchk", " pr", " printenv", " printf", " ptx", " pwd", " readlink", " realpath", " rm", " rmdir", " runcon", " seq", " sha1sum", " sha224sum", " sha256sum", " sha384sum", " sha512sum", " shred", " shuf", " sleep", " sort", " split", " stat", " stdbuf", " stty", " sum", " sync", " tac", " tail", " tee", " test", " timeout", " touch", " tr", " true", " truncate", " tsort", " tty", " uname", " unexpand", " uniq", " unlink", " vdir", " wc", " whoami", " yes", "corkscrew", " corkscrew", "cpio", " cpio", "cppi", " cppi", "cscope", " cscope", " ocs", "ctags", " ctags", " readtags", "curl", " curl", "curseofwar", " curseofwar", "cvs", " cvs", " rcs2log", "daemonize", " daemonize", "darkhttpd", " darkhttpd", "dash", " dash", " sh", "datamash", " datamash", "db", " db_archive", " db_checkpoint", " db_convert", " db_deadlock", " db_dump", " db_hotbackup", " db_load", " db_log_verify", " db_printlog", " db_recover", " db_replicate", " db_stat", " db_tuner", " db_upgrade", " db_verify", "dcraw", " dcraw", "ddrescue", " ddrescue", " ddrescuelog", "debianutils", " add-shell", " ischroot", " remove-shell", " run-parts", " savelog", " tempfile", " which", "dialog", " dialog", " whiptail", "diffutils", " cmp", " diff", " diff3", " sdiff", "direvent", " direvent", "dirmngr", " dirmngr", " dirmngr-client", "dnsutils", " dig", " host", " nslookup", " nsupdate", "dos2unix", " dos2unix", " mac2unix", " unix2dos", " unix2mac", "dpkg", " dpkg", " dpkg-deb", " dpkg-divert", " dpkg-genbuildinfo", " dpkg-query", " dpkg-split", " dpkg-trigger", "dropbear", " dbclient", " dropbear", " dropbearconvert", " dropbearkey", " dropbearmulti", "dvtm", " dvtm", " dvtm-status", "ed", " ed", " red", "elfutils", " eu-addr2line", " eu-elfcmp", " eu-elfcompress", " eu-elflint", " eu-findtextrel", " eu-make-debug-archive", " eu-nm", " eu-objdump", " eu-ranlib", " eu-readelf", " eu-size", " eu-stack", " eu-strings", " eu-strip", " eu-unstrip", "elinks", " elinks", "emacs", " ebrowse", " emacs", " emacs-25.3", " emacsclient", " etags", "erlang", " ct_run", " dialyzer", " epmd", " erl", " erlc", " escript", " run_erl", " to_erl", "espeak", " espeak", "expect", " autoexpect", " expect", " timed-read", " timed-run", " unbuffer", "fdupes", " fdupes", "ffmpeg", " ffmpeg", " ffprobe", "fftw-dev", " fftw-wisdom", " fftw-wisdom-to-conf", " fftwf-wisdom", " fftwl-wisdom", "figlet", " chkfont", " figlet", " figlist", " showfigfonts", "file", " file", "finch", " finch", "findutils", " find", " xargs", "fish", " column", " fish", " fish_indent", " fish_key_reader", "flac", " flac", " metaflac", "flex", " flex", " flex++", "fontconfig-utils", " fc-cache", " fc-cat", " fc-list", " fc-match", " fc-pattern", " fc-query", " fc-scan", " fc-validate", "fortune", " fortune", "fossil", " fossil", "freetype-dev", " freetype-config", "frobtads", " frob", " t3make", " tadsc", "frotz", " frotz", " zgames", "fsmon", " fsmon", "fwknop", " fwknop", "fzf", " fzf", " fzf-tmux", "gawk", " awk", " gawk", "gbt", " gbt", "gcal", " gcal", " gcal2txt", " tcal", " txt2gcal", "gdb", " gcore", " gdb", " gdbserver", "gdbm", " gdbm_dump", " gdbm_load", " gdbmtool", "gdk-pixbuf", " gdk-pixbuf-csource", " gdk-pixbuf-pixdata", " gdk-pixbuf-query-loaders", "gegl", " gcut", " gegl", " gegl-imgcmp", "getconf", " getconf", "gettext", " autopoint", " envsubst", " gettext", " gettext.sh", " gettextize", " msgattrib", " msgcat", " msgcmp", " msgcomm", " msgconv", " msgen", " msgexec", " msgfilter", " msgfmt", " msggrep", " msginit", " msgmerge", " msgunfmt", " msguniq", " ngettext", " recode-sr-latin", " xgettext", "ghostscript", " dvipdf", " eps2eps", " gs", " gsbj", " gsdj", " gsdj500", " gslj", " gslp", " gsnd", " lprsetup.sh", " pdf2dsc", " pdf2ps", " pf2afm", " pfbtopfa", " pphs", " printafm", " ps2ascii", " ps2epsi", " ps2pdf", " ps2pdf12", " ps2pdf13", " ps2pdf14", " ps2pdfwr", " ps2ps", " ps2ps2", " unix-lpr.sh", "gifsicle", " gifdiff", " gifsicle", "git", " git", " git-receive-pack", " git-upload-archive", " git-upload-pack", "glib-bin", " gapplication", " gdbus", " gio", " gio-querymodules", " glib-compile-resources", " glib-compile-schemas", " glib-genmarshal", " glib-mkenums", " gobject-query", " gresource", " gsettings", " gtester", "global", " global", " globash", " gozilla", " gtags", " gtags-cscope", " htags", " htags-server", "glulxe", " glulxe", "gmic", " gmic", "gnuchess", " gnuchess", "gnugo", " gnugo", "gnuit", " .gitaction", " gitaction", " gitdpkgname", " gitfm", " gitkeys", " gitmkdirs", " gitmount", " gitps", " gitregrep", " gitrfgrep", " gitrgrep", " gitunpack", " gitview", " gitwhich", " gitwipe", " gitxgrep", "gnupg", " gpg", " gpg-zip", " gpgsplit", "gnupg2", " addgnupghome", " applygnupgdefaults", " gpg-agent", " gpg-connect-agent", " gpg2", " gpgconf", " gpgparsemail", " gpgscm", " gpgsm", " gpgtar", " gpgv2", " kbxutil", " watchgnupg", "gnuplot", " gnuplot", "gnushogi", " gnushogi", "gnutls", " certtool", " gnutls-cli", " gnutls-cli-debug", " gnutls-serv", " ocsptool", " psktool", " srptool", "golang", " go", " gofmt", "gperf", " gperf", "gpgme", " gpgme-tool", "gpgme-dev", " gpgme-config", "gpgv", " gpgv", "gpsbabel", " gpsbabel", "graphicsmagick", " gm", "graphviz", " acyclic", " bcomps", " ccomps", " circo", " cluster", " diffimg", " dijkstra", " dot", " dot2gxl", " dot_builtins", " edgepaint", " fdp", " gc", " gml2gv", " graphml2gv", " gv2gml", " gv2gxl", " gvcolor", " gvgen", " gvmap", " gvmap.sh", " gvpack", " gvpr", " gxl2dot", " gxl2gv", " mm2gv", " neato", " nop", " osage", " patchwork", " prune", " sccmap", " sfdp", " tred", " twopi", " unflatten", "greed", " greed", "grep", " egrep", " fgrep", " grep", "gst-plugins-base", " gst-device-monitor-1.0", " gst-discoverer-1.0", " gst-play-1.0", "gstreamer", " gst-inspect-1.0", " gst-launch-1.0", " gst-stats-1.0", " gst-typefind-1.0", "gtypist", " gtypist", "gzip", " gunzip", " gzexe", " gzip", " uncompress", " zcat", " zcmp", " zdiff", " zegrep", " zfgrep", " zforce", " zgrep", " zless", " zmore", " znew", "harfbuzz-utils", " hb-ot-shape-closure", " hb-shape", " hb-view", "hashdeep", " hashdeep", " md5deep", " sha1deep", " sha256deep", " tigerdeep", " whirlpooldeep", "hexcurse", " hexcurse", "heyu", " heyu", "hfsutils", " hattrib", " hcd", " hcopy", " hdel", " hdir", " hformat", " hfsutil", " hls", " hmkdir", " hmount", " hpwd", " hrename", " hrmdir", " humount", " hvol", "htop", " htop", "httping", " httping", "hub", " hub", "hunspell", " affixcompress", " analyze", " chmorph", " hunspell", " hunzip", " hzip", " ispellaff2myspell", " makealias", " munch", " unmunch", " wordforms", " wordlist2hunspell", "hydra", " dpl4hydra.sh", " hydra", " hydra-wizard.sh", " pw-inspector", "icecast", " icecast", "iconv", " iconv", "icu-devtools", " derb", " escapesrc", " genbrk", " genccode", " gencfu", " gencmn", " gencnval", " gendict", " gennorm2", " genrb", " gensprep", " icu-config", " icuinfo", " icupkg", " makeconv", " pkgdata", " uconv", "imagemagick", " animate", " compare", " composite", " conjure", " convert", " display", " identify", " import", " magick", " magick-script", " mogrify", " montage", " stream", "imgflo", " imgflo", " imgflo-graphinfo", " imgflo-runtime", "inetutils", " dnsdomainname", " ftp", " logger", " telnet", " tftp", " whois", "inotify-tools", " inotifywait", " inotifywatch", "iperf3", " iperf3", "ired", " bdiff", " ired", " vired", "irssi", " irssi", "jhead", " jhead", "joe", " jmacs", " joe", " jpico", " jstar", " rjoe", "jq", " jq", "jupp", " jmacs", " joe", " jpico", " jstar", " jupp", " rjoe", "kona", " k", "krb5", " compile_et", " gss-client", " gss-server", " k5srvutil", " kadmin", " kadmin.local", " kadmind", " kdb5_util", " kdestroy", " kinit", " klist", " kpasswd", " kprop", " kpropd", " kproplog", " krb5-config", " krb5-send-pr", " krb5kdc", " ksu", " kswitch", " ktutil", " kvno", " sclient", " sim_client", " sim_server", " sserver", " uuclient", " uuserver", "ldc", " dub", " ldc-build-runtime", " ldc2", " ldmd2", " rdmd", "ldns-dev", " ldns-config", "ledger", " ledger", "leptonica", " convertfilestopdf", " convertfilestops", " convertformat", " convertsegfilestopdf", " convertsegfilestops", " converttopdf", " converttops", " fileinfo", " xtractprotos", "less", " less", " lessecho", " lesskey", "lftp", " lftp", " lftpget", "libassuan-dev", " libassuan-config", "libcaca", " caca-config", " cacaclock", " cacademo", " cacafire", " cacaplay", " cacaserver", " cacaview", " img2txt", "libcairo", " cairo-trace", "libcroco", " croco-0.6-config", " csslint-0.6", "libcurl-dev", " curl-config", "libgcrypt", " dumpsexp", " hmac256", " libgcrypt-config", " mpicalc", "libgd", " annotate", " bdftogd", " gd2copypal", " gd2togif", " gd2topng", " gdcmpgif", " gdlib-config", " gdparttopng", " gdtopng", " giftogd2", " pngtogd", " pngtogd2", " webpng", "libgpg-error", " gpg-error", " gpg-error-config", "libgrpc", " check_epollexclusive", " gen_hpack_tables", " gen_legal_metadata_characters", " gen_percent_encoding_tables", " grpc_cpp_plugin", " grpc_create_jwt", " grpc_csharp_plugin", " grpc_node_plugin", " grpc_objective_c_plugin", " grpc_php_plugin", " grpc_print_google_default_creds_token", " grpc_python_plugin", " grpc_ruby_plugin", " grpc_verify_jwt", "libjasper-utils", " imgcmp", " imginfo", " jasper", "libjpeg-turbo-progs", " cjpeg", " djpeg", " jpegtran", " rdjpgcom", " tjbench", " wrjpgcom", "libksba-dev", " ksba-config", "libnpth", " npth-config", "liboggz", " oggz", " oggz-chop", " oggz-codecs", " oggz-comment", " oggz-diff", " oggz-dump", " oggz-info", " oggz-known-codecs", " oggz-merge", " oggz-rip", " oggz-scan", " oggz-sort", " oggz-validate", "libotr", " otr_mackey", " otr_modify", " otr_parse", " otr_readforge", " otr_remac", " otr_sesskeys", "libpng-dev", " libpng-config", " libpng16-config", "librsvg", " rsvg-convert", "libtiff-utils", " fax2ps", " fax2tiff", " pal2rgb", " ppm2tiff", " raw2tiff", " tiff2bw", " tiff2pdf", " tiff2ps", " tiff2rgba", " tiffcmp", " tiffcp", " tiffcrop", " tiffdither", " tiffdump", " tiffinfo", " tiffmedian", " tiffset", " tiffsplit", "libtool", " libtool", " libtoolize", "libxapian-dev", " xapian-config", "libxml2-dev", " xml2-config", "libxml2-utils", " xmlcatalog", " xmllint", "libxslt-dev", " xslt-config", "libzmq", " curve_keygen", "lighttpd", " lighttpd", "littlecms-utils", " jpgicc", " linkicc", " psicc", " tificc", " transicc", "lld", " ld.lld", " lld", " lld-link", "llvm", " llc", " lli", " llvm-ar", " llvm-as", " llvm-bcanalyzer", " llvm-c-test", " llvm-cat", " llvm-config", " llvm-cov", " llvm-cvtres", " llvm-cxxdump", " llvm-cxxfilt", " llvm-diff", " llvm-dis", " llvm-dlltool", " llvm-dsymutil", " llvm-dwarfdump", " llvm-dwp", " llvm-extract", " llvm-lib", " llvm-link", " llvm-lto", " llvm-lto2", " llvm-mc", " llvm-mcmarkup", " llvm-modextract", " llvm-mt", " llvm-nm", " llvm-objdump", " llvm-opt-report", " llvm-pdbutil", " llvm-profdata", " llvm-ranlib", " llvm-readelf", " llvm-readobj", " llvm-rtdyld", " llvm-size", " llvm-split", " llvm-stress", " llvm-strings", " llvm-symbolizer", " llvm-xray", " obj2yaml", " opt", " verify-uselistorder", " yaml2obj", "ltrace", " ltrace", "lua", " lua", " luac", "lynx", " lynx", "lzip", " lzip", "lzop", " lzop", "m4", " m4", "macchanger", " macchanger", "make", " make", "man", " apropos", " demandoc", " makewhatis", " man", " mandoc", " soelim", " whatis", "mariadb", " aria_chk", " aria_dump_log", " aria_ftdump", " aria_pack", " aria_read_log", " innochecksum", " mariadb_config", " msql2mysql", " my_print_defaults", " myisam_ftdump", " myisamchk", " myisamlog", " myisampack", " mysql", " mysql_client_test", " mysql_config", " mysql_convert_table_format", " mysql_find_rows", " mysql_fix_extensions", " mysql_install_db", " mysql_plugin", " mysql_secure_installation", " mysql_setpermission", " mysql_tzinfo_to_sql", " mysql_upgrade", " mysql_waitpid", " mysqlaccess", " mysqladmin", " mysqlbinlog", " mysqlcheck", " mysqld", " mysqld_multi", " mysqld_safe", " mysqld_safe_helper", " mysqldump", " mysqldumpslow", " mysqlhotcopy", " mysqlimport", " mysqlshow", " mysqlslap", " mytop", " perror", " replace", " resolve_stack_dump", " resolveip", "mathomatic", " mathomatic", "mc", " mc", " mcdiff", " mcedit", " mcview", "megatools", " megacopy", " megadf", " megadl", " megaget", " megals", " megamkdir", " megaput", " megareg", " megarm", "memcached", " memcached", "micro", " micro", "minicom", " ascii-xfr", " minicom", " runscript", " xminicom", "mlocate", " locate", " updatedb", "moon-buggy", " moon-buggy", "moria", " moria", "mosh", " mosh", " mosh-client", " mosh-server", " mosh.pl", "mosquitto", " mosquitto", " mosquitto_passwd", " mosquitto_pub", " mosquitto_sub", "mp3splt", " mp3splt", "mpc", " mpc", "mpd", " mpd", "mpv", " mpv", "mtools", " amuFormat.sh", " lz", " mattrib", " mbadblocks", " mcat", " mcd", " mcheck", " mclasserase", " mcomp", " mcopy", " mdel", " mdeltree", " mdir", " mdu", " mformat", " minfo", " mkmanifest", " mlabel", " mmd", " mmount", " mmove", " mpartition", " mrd", " mren", " mshortname", " mshowfat", " mtools", " mtoolstest", " mtype", " mxtar", " mzip", " tgz", " uz", "multitail", " multitail", "mutt", " mutt", " pgpewrap", " pgpring", " smime_keys", "nano", " nano", "ncdc", " ncdc", "ncdu", " ncdu", "ncmpcpp", " ncmpcpp", "ncurses-dev", " ncursesw6-config", "ncurses-utils", " clear", " infocmp", " reset", " tabs", " tic", " toe", " tput", " tset", "ne", " ne", "neovim", " nvim", "net-tools", " arp", " ifconfig", " ipmaddr", " iptunnel", " mii-tool", " nameif", " netstat", " plipconfig", " rarp", " route", " slattach", "netcat", " nc", " ncat", " netcat", "netpbm", " 411toppm", " anytopnm", " asciitopgm", " atktopbm", " avstopam", " bioradtopgm", " bmptopnm", " bmptoppm", " brushtopbm", " cameratopam", " cistopbm", " cmuwmtopbm", " ddbugtopbm", " escp2topbm", " eyuvtoppm", " fiascotopnm", " fitstopnm", " fstopgm", " g3topbm", " gemtopbm", " gemtopnm", " giftopnm", " gouldtoppm", " hdifftopam", " hipstopgm", " hpcdtoppm", " icontopbm", " ilbmtoppm", " imgtoppm", " infotopam", " jbigtopnm", " jpeg2ktopam", " jpegtopnm", " leaftoppm", " lispmtopgm", " macptopbm", " manweb", " mdatopbm", " mgrtopbm", " mrftopbm", " mtvtoppm", " neotoppm", " netpbm-config", " palmtopnm", " pamaddnoise", " pamarith", " pambackground", " pambayer", " pamchannel", " pamcomp", " pamcrater", " pamcut", " pamdeinterlace", " pamdepth", " pamdice", " pamditherbw", " pamedge", " pamendian", " pamenlarge", " pamexec", " pamfile", " pamfix", " pamfixtrunc", " pamflip", " pamfunc", " pamgauss", " pamgradient", " pamlookup", " pammasksharpen", " pammixinterlace", " pammosaicknit", " pamoil", " pampaintspill", " pamperspective", " pampick", " pampop9", " pamrecolor", " pamrgbatopng", " pamrubber", " pamscale", " pamseq", " pamshadedrelief", " pamsharpmap", " pamsharpness", " pamsistoaglyph", " pamslice", " pamsplit", " pamstack", " pamstereogram", " pamstretch", " pamstretch-gen", " pamsumm", " pamsummcol", " pamtable", " pamthreshold", " pamtilt", " pamtoavs", " pamtodjvurle", " pamtofits", " pamtogif", " pamtohdiff", " pamtohtmltbl", " pamtojpeg2k", " pamtompfont", " pamtooctaveimg", " pamtopam", " pamtopdbimg", " pamtopfm", " pamtopng", " pamtopnm", " pamtosrf", " pamtosvg", " pamtotga", " pamtotiff", " pamtouil", " pamtowinicon", " pamtoxvmini", " pamundice", " pamunlookup", " pamvalidate", " pamwipeout", " pbmclean", " pbmlife", " pbmmake", " pbmmask", " pbmminkowski", " pbmpage", " pbmpscale", " pbmreduce", " pbmtext", " pbmtextps", " pbmto10x", " pbmto4425", " pbmtoascii", " pbmtoatk", " pbmtobbnbg", " pbmtocis", " pbmtocmuwm", " pbmtodjvurle", " pbmtoepsi", " pbmtoepson", " pbmtoescp2", " pbmtog3", " pbmtogem", " pbmtogo", " pbmtoibm23xx", " pbmtoicon", " pbmtolj", " pbmtoln03", " pbmtolps", " pbmtomacp", " pbmtomatrixorbital", " pbmtomda", " pbmtomgr", " pbmtomrf", " pbmtonokia", " pbmtopgm", " pbmtopi3", " pbmtopk", " pbmtoplot", " pbmtoppa", " pbmtopsg3", " pbmtoptx", " pbmtosunicon", " pbmtowbmp", " pbmtox10bm", " pbmtoxbm", " pbmtoybm", " pbmtozinc", " pbmupc", " pc1toppm", " pcdovtoppm", " pcxtoppm", " pdbimgtopam", " pfmtopam", " pgmabel", " pgmbentley", " pgmcrater", " pgmdeshadow", " pgmedge", " pgmenhance", " pgmhist", " pgmkernel", " pgmmake", " pgmmedian", " pgmminkowski", " pgmmorphconv", " pgmnoise", " pgmnorm", " pgmoil", " pgmramp", " pgmslice", " pgmtexture", " pgmtofs", " pgmtolispm", " pgmtopbm", " pgmtopgm", " pgmtoppm", " pgmtosbig", " pgmtost4", " pi1toppm", " pi3topbm", " picttoppm", " pjtoppm", " pktopbm", " pngtopam", " pngtopnm", " pnmalias", " pnmarith", " pnmcat", " pnmcolormap", " pnmcomp", " pnmconvol", " pnmcrop", " pnmcut", " pnmdepth", " pnmenlarge", " pnmfile", " pnmflip", " pnmgamma", " pnmhisteq", " pnmhistmap", " pnmindex", " pnminterp", " pnminvert", " pnmmargin", " pnmmercator", " pnmmontage", " pnmnlfilt", " pnmnoraw", " pnmnorm", " pnmpad", " pnmpaste", " pnmpsnr", " pnmquant", " pnmquantall", " pnmremap", " pnmrotate", " pnmscale", " pnmscalefixed", " pnmshear", " pnmsmooth", " pnmsplit", " pnmstitch", " pnmtile", " pnmtoddif", " pnmtofiasco", " pnmtofits", " pnmtojbig", " pnmtojpeg", " pnmtopalm", " pnmtopclxl", " pnmtoplainpnm", " pnmtopng", " pnmtopnm", " pnmtops", " pnmtorast", " pnmtorle", " pnmtosgi", " pnmtosir", " pnmtotiff", " pnmtotiffcmyk", " pnmtoxwd", " ppm3d", " ppmbrighten", " ppmchange", " ppmcie", " ppmcolormask", " ppmcolors", " ppmdcfont", " ppmddumpfont", " ppmdim", " ppmdist", " ppmdither", " ppmdmkfont", " ppmdraw", " ppmfade", " ppmflash", " ppmforge", " ppmglobe", " ppmhist", " ppmlabel", " ppmmake", " ppmmix", " ppmnorm", " ppmntsc", " ppmpat", " ppmquant", " ppmquantall", " ppmrainbow", " ppmrelief", " ppmrough", " ppmshadow", " ppmshift", " ppmspread", " ppmtoacad", " ppmtoapplevol", " ppmtoarbtxt", " ppmtoascii", " ppmtobmp", " ppmtoeyuv", " ppmtogif", " ppmtoicr", " ppmtoilbm", " ppmtojpeg", " ppmtoleaf", " ppmtolj", " ppmtomap", " ppmtomitsu", " ppmtompeg", " ppmtoneo", " ppmtopcx", " ppmtopgm", " ppmtopi1", " ppmtopict", " ppmtopj", " ppmtopjxl", " ppmtoppm", " ppmtopuzz", " ppmtorgb3", " ppmtosixel", " ppmtospu", " ppmtoterm", " ppmtotga", " ppmtouil", " ppmtowinicon", " ppmtoxpm", " ppmtoyuv", " ppmtoyuvsplit", " ppmtv", " ppmwheel", " psidtopgm", " pstopnm", " qrttoppm", " rasttopnm", " rawtopgm", " rawtoppm", " rgb3toppm", " rlatopam", " rletopnm", " sbigtopgm", " sgitopnm", " sirtopnm", " sldtoppm", " spctoppm", " spottopgm", " sputoppm", " srftopam", " st4topgm", " sunicontopnm", " svgtopam", " tgatoppm", " thinkjettopbm", " tifftopnm", " wbmptopbm", " winicontopam", " winicontoppm", " xbmtopbm", " ximtoppm", " xpmtoppm", " xvminitoppm", " xwdtopnm", " ybmtopbm", " yuvsplittoppm", " yuvtoppm", " yuy2topam", " zeisstopnm", "nettle", " nettle-hash", " nettle-lfib-stream", " nettle-pbkdf2", " pkcs1-conv", " sexp-conv", "newsboat", " newsboat", " podboat", "nginx", " nginx", "ninja", " ninja", "nmap", " nmap", " nping", "nodejs", " node", " npm", " npx", "nodejs-current", " node", " npm", " npx", "notmuch", " notmuch", "nyancat", " nyancat", "nzbget", " nzbget", "oathtool", " oathtool", " pskctool", "ocrad", " ocrad", "openjpeg-tools", " opj_compress", " opj_decompress", " opj_dump", "openssh", " scp", " sftp", " source-ssh-agent", " ssh", " ssh-add", " ssh-agent", " ssh-copy-id", " ssh-keygen", " ssh-keyscan", " ssha", " sshd", "openssl-tool", " openssl", "optipng", " optipng", "opus-tools", " opusdec", " opusenc", " opusinfo", "p7zip", " 7z", " 7za", " 7zr", "pango", " pango-view", "par2", " par2", " par2create", " par2repair", " par2verify", "patch", " patch", "patchelf", " patchelf", "pcre-dev", " pcre-config", "pcre2-dev", " pcre2-config", "perl", " corelist", " cpan", " enc2xs", " encguess", " h2ph", " h2xs", " instmodsh", " json_pp", " libnetcfg", " perl", " perlbug", " perldoc", " perlivp", " perlthanks", " piconv", " pl2pm", " pod2html", " pod2man", " pod2text", " pod2usage", " podchecker", " podselect", " prove", " ptar", " ptardiff", " ptargrep", " shasum", " splain", " xsubpp", " zipdetails", "pforth", " pforth", "php", " phar", " phar.phar", " php", " php-cgi", " php-config", " phpdbg", " phpize", "php-fpm", " php-fpm", "pick", " pick", "picolisp", " picolisp", " pil", "pinentry", " pinentry", " pinentry-curses", "pkg-config", " pkg-config", "play-audio", " play-audio", "pngquant", " pngquant", "poppler", " pdfdetach", " pdffonts", " pdfimages", " pdfinfo", " pdfseparate", " pdftocairo", " pdftohtml", " pdftoppm", " pdftops", " pdftotext", " pdfunite", "postgresql", " clusterdb", " createdb", " createuser", " dropdb", " dropuser", " initdb", " pg_archivecleanup", " pg_basebackup", " pg_config", " pg_controldata", " pg_ctl", " pg_dump", " pg_dumpall", " pg_isready", " pg_receivewal", " pg_recvlogical", " pg_resetwal", " pg_restore", " pg_rewind", " pg_test_fsync", " pg_test_timing", " pg_upgrade", " pg_waldump", " pgbench", " postgres", " postmaster", " psql", " reindexdb", " vacuumdb", "potrace", " mkbitmap", " potrace", "privoxy", " privoxy", "procps", " free", " pgrep", " pidof", " pkill", " pmap", " ps", " pwdx", " slabtop", " sysctl", " tload", " top", " uptime", " vmstat", " w", " watch", "profanity", " profanity", "proot", " proot", " termux-chroot", "protobuf", " protoc", "psmisc", " fuser", " killall", " peekfd", " prtstat", " pstree", "pulseaudio", " esdcompat", " pacat", " pacmd", " pactl", " pasuspender", " pulseaudio", "pure-ftpd", " pure-authd", " pure-ftpd", " pure-ftpwho", " pure-mrtginfo", " pure-pw", " pure-pwconvert", " pure-quotacheck", " pure-statsdecode", " pure-uploadscript", "pv", " pv", "pwgen", " pwgen", "python", " 2to3", " 2to3-3.6", " pydoc3", " pydoc3.6", " python", " python3", " python3-config", " python3.6", " python3.6-config", " python3.6m-config", " pyvenv", " pyvenv-3.6", "python2", " idle", " pydoc", " python-config", " python2", " python2-config", " python2.7", " python2.7-config", " smtpd.py", "qalc", " qalc", "qpdf", " fix-qdf", " qpdf", " zlib-flate", "radare2", " r2", " r2agent", " r2pm", " rabin2", " radare2", " radiff2", " rafind2", " ragg2", " ragg2-cc", " rahash2", " rarun2", " rasm2", " rax2", "ragel", " ragel", "rcs", " ci", " co", " ident", " merge", " rcs", " rcsclean", " rcsdiff", " rcsmerge", " rlog", "rdiff", " rdiff", "redir", " redir", "redis", " redis-benchmark", " redis-check-aof", " redis-check-rdb", " redis-cli", " redis-sentinel", " redis-server", "remind", " rem", " rem2ps", " remind", "rgbds", " rgbasm", " rgbfix", " rgbgfx", " rgblink", "rlwrap", " rlwrap", "rsync", " rsync", "rtmpdump", " rtmpdump", " rtmpgw", " rtmpsrv", " rtmpsuck", "ruby", " erb", " gem", " irb", " rake", " rdoc", " ruby", "screen", " screen", " screen-4.6.2", "scrypt", " scrypt", "sed", " sed", "sensible-utils", " select-editor", " sensible-browser", " sensible-editor", " sensible-pager", "sharutils", " shar", " unshar", " uudecode", " uuencode", "silversearcher-ag", " ag", "sl", " sl", "socat", " filan", " procan", " socat", "sox", " play", " rec", " sox", " soxi", "sqlite", " sqlite3", "squid", " purge", " squid", " squidclient", "sshpass", " sshpass", "sslscan", " sslscan", "stag", " stag", "strace", " strace", " strace-log-merge", "stunnel", " stunnel", "subversion", " svn", " svnadmin", " svnbench", " svndumpfilter", " svnfsfs", " svnlook", " svnmucc", " svnrdump", " svnserve", " svnsync", " svnversion", "syncthing", " syncthing", "tar", " tar", "tasksh", " tasksh", "taskwarrior", " task", "tcl", " sqlite3_analyzer", " tclsh", " tclsh8.6", "tcsh", " tcsh", "teckit", " sfconv", " teckit_compile", " txtconv", "termux-api", " termux-audio-info", " termux-battery-status", " termux-camera-info", " termux-camera-photo", " termux-clipboard-get", " termux-clipboard-set", " termux-contact-list", " termux-dialog", " termux-download", " termux-infrared-frequencies", " termux-infrared-transmit", " termux-location", " termux-media-scan", " termux-notification", " termux-notification-remove", " termux-share", " termux-sms-inbox", " termux-sms-send", " termux-storage-get", " termux-telephony-call", " termux-telephony-cellinfo", " termux-telephony-deviceinfo", " termux-toast", " termux-tts-engines", " termux-tts-speak", " termux-vibrate", " termux-wifi-connectioninfo", " termux-wifi-scaninfo", "termux-elf-cleaner", " termux-elf-cleaner", "teseq", " reseq", " teseq", "tesseract", " tesseract", "texinfo", " info", " install-info", " makeinfo", " pdftexi2dvi", " pod2texi", " texi2any", " texi2dvi", " texi2pdf", " texindex", "texlive-bin", " a2ping", " a5toa4", " adhocfilelist", " afm2afm", " afm2pl", " afm2tfm", " aleph", " allcm", " allec", " allneeded", " arara", " arlatex", " authorindex", " autoinst", " autosp", " bbl2bib", " bbox", " bg5+latex", " bg5+pdflatex", " bg5conv", " bg5latex", " bg5pdflatex", " bibdoiadd", " bibexport", " bibmradd", " bibtex", " bibtex8", " bibzbladd", " bundledoc", " cachepic", " cef5conv", " cef5latex", " cef5pdflatex", " cefconv", " ceflatex", " cefpdflatex", " cefsconv", " cefslatex", " cefspdflatex", " cfftot1", " checkcites", " checklistings", " chktex", " chkweb", " cjk-gs-integrate", " context", " contextjit", " convbkmk", " ctangle", " ctanify", " ctanupload", " ctie", " cweave", " de-macro", " depythontex", " detex", " devnag", " deweb", " diadia", " disdvi", " dosepsbin", " dt2dv", " dtxgen", " dv2dt", " dvi2fax", " dvi2tty", " dviasm", " dvicopy", " dvidvi", " dvigif", " dvihp", " dvilj", " dvilj2p", " dvilj4", " dvilj4l", " dvilj6", " dvipdfm", " dvipdfmx", " dvipdft", " dvipng", " dvipos", " dvips", " dvired", " dvitomp", " dvitype", " e2pall", " ebb", " ebong", " epspdf", " epspdftk", " epstopdf", " eptex", " euptex", " exceltex", " extconv", " extractbb", " fig4latex", " findhyph", " fmtutil", " fmtutil-sys", " fmtutil-user", " fontinst", " fragmaster", " gbklatex", " gbkpdflatex", " getmapdl", " gftodvi", " gftopk", " gftype", " gsftopk", " hbf2gf", " ht", " htcontext", " htlatex", " htmex", " httex", " httexi", " htxelatex", " htxetex", " inimf", " initex", " installfont-tl", " jamo-normalize", " kanji-config-updmap", " kanji-config-updmap-sys", " kanji-config-updmap-user", " kanji-fontmap-creator", " komkindex", " kpseaccess", " kpsepath", " kpsereadlink", " kpsestat", " kpsetool", " kpsewhere", " kpsewhich", " kpsexpand", " lacheck", " latex-git-log", " latex-papersize", " latex2man", " latex2nemeth", " latexdiff", " latexdiff-vc", " latexfileversion", " latexindent", " latexmk", " latexpand", " latexrevise", " lily-glyph-commands", " lily-image-commands", " lily-rebuild-pdfs", " listbib", " listings-ext.sh", " ltx2crossrefxml", " ltxfileinfo", " ltximg", " lua2dox_filter", " luaotfload-tool", " luatex", " luatools", " lwarpmk", " m-tx", " make4ht", " makedtx", " makeglossaries", " makeglossaries-lite", " makeindex", " match_parens", " mathspic", " mf", " mf-nowin", " mf2pt1", " mfplain", " mft", " mk4ht", " mkgrkindex", " mkindex", " mkjobtexmf", " mkocp", " mkofm", " mkpic", " mkt1font", " mktexfmt", " mktexlsr", " mktexmf", " mktexpk", " mktextfm", " mmafm", " mmpfb", " mpost", " mptopdf", " mtxrun", " mtxrunjit", " multibibliography", " musixflx", " musixtex", " odvicopy", " odvitype", " ofm2opl", " omfonts", " opl2ofm", " ot2kpx", " otangle", " otfinfo", " otftotfm", " otp2ocp", " outocp", " ovf2ovp", " ovp2ovf", " patgen", " pbibtex", " pdf180", " pdf270", " pdf90", " pdfannotextractor", " pdfatfi", " pdfbook", " pdfbook2", " pdfcrop", " pdfflip", " pdfjam", " pdfjam-pocketmod", " pdfjam-slides3up", " pdfjam-slides6up", " pdfjoin", " pdflatexpicscale", " pdfnup", " pdfpun", " pdftex", " pdftosrc", " pdfxup", " pdvitomp", " pdvitype", " pedigree", " perltex", " pfarrei", " pkfix", " pkfix-helper", " pktogf", " pktype", " pltotf", " pmpost", " pmxchords", " pn2pdf", " pooltype", " ppltotf", " prepmx", " ps2eps", " ps2frag", " ps4pdf", " pslatex", " pst2pdf", " ptex", " ptex2pdf", " ptftopl", " purifyeps", " pygmentex", " pythontex", " repstopdf", " rpdfcrop", " rubibtex", " rubikrotation", " rumakeindex", " rungs", " simpdftex", " sjisconv", " sjislatex", " sjispdflatex", " splitindex", " srcredact", " sty2dtx", " svn-multi", " synctex", " t1dotlessj", " t1lint", " t1rawafm", " t1reencode", " t1testpage", " t4ht", " tangle", " tex", " tex4ebook", " tex4ht", " texconfig", " texconfig-dialog", " texconfig-sys", " texcount", " texdef", " texdiff", " texdirflatten", " texdoc", " texdoctk", " texexec", " texfot", " texhash", " texlinks", " texliveonfly", " texloganalyser", " texlua", " texluac", " texmfstart", " texosquery", " texosquery-jre5", " texosquery-jre8", " tftopl", " thumbpdf", " tie", " tlmgr", " tlmgr.ln", " tpic2pdftex", " ttf2afm", " ttf2kotexfont", " ttf2pk", " ttf2tfm", " ttftotype42", " typeoutfileinfo", " ulqda", " upbibtex", " updmap", " updmap-sys", " updmap-user", " updvitomp", " updvitype", " upmendex", " upmpost", " uppltotf", " uptex", " uptftopl", " urlbst", " vftovp", " vlna", " vpe", " vpl2ovp", " vpl2vpl", " vptovf", " weave", " wofm2opl", " wopl2ofm", " wovf2ovp", " wovp2ovf", " xdvipdfmx", " xetex", " xhlatex", " yplan", "tig", " tig", "timewarrior", " timew", "tintin++", " tt++", "tinyscheme", " tinyscheme", "tmate", " tmate", "tmux", " tmux", "toilet", " toilet", "tor", " tor", " tor-gencert", " tor-resolve", " torify", "torsocks", " torsocks", "tracepath", " tracepath", " traceroute", "transmission", " transmission-create", " transmission-daemon", " transmission-edit", " transmission-remote", " transmission-show", "tree", " tree", "tsocks", " tsocks", "tty-clock", " tty-clock", "tty-solitaire", " ttysolitaire", "ttyrec", " ttyplay", " ttyrec", " ttytime", "units", " units", " units_cur", "unrar", " unrar", "unzip", " funzip", " unzip", " unzipsfx", " zipgrep", " zipinfo", "utfdecode", " utfdecode", "util-linux", " addpart", " blkdiscard", " blkid", " blkzone", " blockdev", " cal", " chcpu", " chrt", " col", " colcrt", " colrm", " ctrlaltdel", " delpart", " dmesg", " fallocate", " fdformat", " fincore", " findfs", " flock", " fsck.cramfs", " fsck.minix", " fsfreeze", " getopt", " hexdump", " hwclock", " ionice", " isosize", " ldattach", " linux32", " linux64", " look", " losetup", " lscpu", " lsipc", " lsns", " mcookie", " mesg", " mkfs", " mkfs.bfs", " mkfs.cramfs", " mkfs.minix", " mkswap", " more", " namei", " nologin", " nsenter", " partx", " prlimit", " raw", " readprofile", " rename", " renice", " resizepart", " rev", " rtcwake", " script", " scriptreplay", " setarch", " setsid", " setterm", " swaplabel", " taskset", " ul", " uname26", " unshare", " wdctl", " whereis", " wipefs", " zramctl", "valac", " vala", " vala-0.38", " vala-gen-introspect", " vala-gen-introspect-0.38", " valac", " valac-0.38", " vapicheck", " vapicheck-0.38", " vapigen", " vapigen-0.38", "valadoc", " valadoc", " valadoc-0.38", "valgrind", " callgrind_annotate", " callgrind_control", " cg_annotate", " cg_diff", " cg_merge", " ms_print", " valgrind", " valgrind-di-server", " valgrind-listener", " vgdb", "vifm", " vifm", " vifm-convert-dircolors", " vifm-pause", " vifm-screen-split", "vim", " vi", " view", " vim", " vimdiff", " vimtutor", " xxd", "vim-python", " vi", " view", " vim", " vimdiff", " vimtutor", " xxd", "vorbis-tools", " oggdec", " oggenc", " ogginfo", " vcut", " vorbiscomment", "vttest", " vttest", "vtutils", " vtquery", " vtshowkeys", " vtsize", " vttitle", "w3m", " w3m", " w3mman", "wcalc", " wcalc", "weechat", " weechat", "wget", " wget", "wol", " wol", " wol-bootptab", " wol-dhcpdconf", "x264", " x264", "x265", " x265", "xapian-tools", " copydatabase", " quest", " simpleexpand", " simpleindex", " simplesearch", " xapian-check", " xapian-compact", " xapian-delve", " xapian-metadata", " xapian-progsrv", " xapian-replicate", " xapian-replicate-server", " xapian-tcpsrv", "xmlsec", " xmlsec1", " xmlsec1-config", "xmlstarlet", " xml", "xorriso", " osirrox", " xorrecord", " xorriso", " xorriso-tcltk", " xorrisofs", "xsltproc", " xsltproc", "xz-utils", " lzcat", " lzcmp", " lzdiff", " lzegrep", " lzfgrep", " lzgrep", " lzless", " lzma", " lzmadec", " lzmainfo", " lzmore", " unlzma", " unxz", " xz", " xzcat", " xzcmp", " xzdec", " xzdiff", " xzegrep", " xzfgrep", " xzgrep", " xzless", " xzmore", "yasm", " vsyasm", " yasm", " ytasm", "zbar", " zbarimg", "zile", " zile", "zip", " zip", " zipcloak", " zipnote", " zipsplit", "zsh", " zsh", "zstd", " unzstd", " zstd", " zstdcat", " zstdgrep", " zstdless", " zstdmt", "ack-grep", " ack", "apksigner", " apksigner", "asciinema", " asciinema", "autoconf", " autoconf", " autoheader", " autom4te", " autoreconf", " autoscan", " autoupdate", " ifnames", "automake", " aclocal", " aclocal-1.15", " automake", " automake-1.15", "byobu", " byobu", " byobu-config", " byobu-ctrl-a", " byobu-disable", " byobu-disable-prompt", " byobu-enable", " byobu-enable-prompt", " byobu-export", " byobu-janitor", " byobu-keybindings", " byobu-launch", " byobu-launcher", " byobu-launcher-install", " byobu-launcher-uninstall", " byobu-layout", " byobu-prompt", " byobu-quiet", " byobu-reconnect-sockets", " byobu-screen", " byobu-select-backend", " byobu-select-profile", " byobu-select-session", " byobu-shell", " byobu-silent", " byobu-status", " byobu-status-detail", " byobu-tmux", " byobu-ugraph", " byobu-ulevel", " col1", " ctail", " manifest", " purge-old-kernels", " vigpg", " wifi-status", "colordiff", " colordiff", "cowsay", " cowsay", "debootstrap", " debootstrap", "dx", " dx", "ecj", " ecj", "getmail", " getmail", " getmail_fetch", " getmail_maildir", " getmail_mbox", "luarocks", " luarocks", " luarocks-5.3", " luarocks-admin", " luarocks-admin-5.3", "neofetch", " neofetch", "parallel", " env_parallel", " env_parallel.ash", " env_parallel.bash", " env_parallel.csh", " env_parallel.dash", " env_parallel.fish", " env_parallel.ksh", " env_parallel.pdksh", " env_parallel.sh", " env_parallel.tcsh", " env_parallel.zsh", " niceload", " parallel", " parcat", " sem", " sql", "pass", " pass", "pastebinit", " pastebinit", "pathpicker", " fpp", "ranger", " ranger", " rifle", "ruby-ri", " ri", "screenfetch", " screenfetch", "stow", " chkstow", " stow", "termux-am", " am", "termux-apt-repo", " termux-apt-repo", "termux-create-package", " termux-create-package", "termux-tools", " chsh", " dalvikvm", " df", " getprop", " ip", " logcat", " login", " ping", " ping6", " pkg", " pm", " settings", " su", " termux-fix-shebang", " termux-info", " termux-open", " termux-open-url", " termux-reload-settings", " termux-setup-storage", " termux-wake-lock", " termux-wake-unlock", " xdg-open", "tsu", " tsu", "vcsh", " vcsh", "yarn", " yarn", }; Desktop version Sign out
Bitcoin: A Peer-to-Peer Electronic Cash System Satoshi Nakamoto satoshin@gmx.com www.bitcoin.org Abstract. A purely peer-to-peer version of electronic cash would allow online payments to be sent directly from one party to another without going through a financial institution. Digital signatures provide part of the solution, but the main benefits are lost if a trusted third party is still required to prevent double-spending. We propose a solution to the double-spending problem using a peer-to-peer network. The network timestamps transactions by hashing them into an ongoing chain of hash-based proof-of-work, forming a record that cannot be changed without redoing the proof-of-work. The longest chain not only serves as proof of the sequence of events witnessed, but proof that it came from the largest pool of CPU power. As long as a majority of CPU power is controlled by nodes that are not cooperating to attack the network, they'll generate the longest chain and outpace attackers. The network itself requires minimal structure. Messages are broadcast on a best effort basis, and nodes can leave and rejoin the network at will, accepting the longest proof-of-work chain as proof of what happened while they were gone. 1. Introduction Commerce on the Internet has come to rely almost exclusively on financial institutions serving as trusted third parties to process electronic payments. While the system works well enough for most transactions, it still suffers from the inherent weaknesses of the trust based model. Completely non-reversible transactions are not really possible, since financial institutions cannot avoid mediating disputes. The cost of mediation increases transaction costs, limiting the minimum practical transaction size and cutting off the possibility for small casual transactions, and there is a broader cost in the loss of ability to make non-reversible payments for non- reversible services. With the possibility of reversal, the need for trust spreads. Merchants must be wary of their customers, hassling them for more information than they would otherwise need. A certain percentage of fraud is accepted as unavoidable. These costs and payment uncertainties can be avoided in person by using physical currency, but no mechanism exists to make payments over a communications channel without a trusted party. What is needed is an electronic payment system based on cryptographic proof instead of trust, allowing any two willing parties to transact directly with each other without the need for a trusted third party. Transactions that are computationally impractical to reverse would protect sellers from fraud, and routine escrow mechanisms could easily be implemented to protect buyers. In this paper, we propose a solution to the double-spending problem using a peer-to-peer distributed timestamp server to generate computational proof of the chronological order of transactions. The system is secure as long as honest nodes collectively control more CPU power than any cooperating group of attacker nodes. 1 2. Transactions We define an electronic coin as a chain of digital signatures. Each owner transfers the coin to the next by digitally signing a hash of the previous transaction and the public key of the next owner and adding these to the end of the coin. A payee can verify the signatures to verify the chain of ownership. Transaction Hash Transaction Hash Transaction Hash Owner 1's Public Key Owner 2's Public Key Owner 3's Public Key Owner 0's Signature Owner 1's Signature The problem of course is the payee can't verify that one of the owners did not double-spend the coin. A common solution is to introduce a trusted central authority, or mint, that checks every transaction for double spending. After each transaction, the coin must be returned to the mint to issue a new coin, and only coins issued directly from the mint are trusted not to be double-spent. The problem with this solution is that the fate of the entire money system depends on the company running the mint, with every transaction having to go through them, just like a bank. We need a way for the payee to know that the previous owners did not sign any earlier transactions. For our purposes, the earliest transaction is the one that counts, so we don't care about later attempts to double-spend. The only way to confirm the absence of a transaction is to be aware of all transactions. In the mint based model, the mint was aware of all transactions and decided which arrived first. To accomplish this without a trusted party, transactions must be publicly announced [1], and we need a system for participants to agree on a single history of the order in which they were received. The payee needs proof that at the time of each transaction, the majority of nodes agreed it was the first received. 3. Timestamp Server The solution we propose begins with a timestamp server. A timestamp server works by taking a hash of a block of items to be timestamped and widely publishing the hash, such as in a newspaper or Usenet post [2-5]. The timestamp proves that the data must have existed at the time, obviously, in order to get into the hash. Each timestamp includes the previous timestamp in its hash, forming a chain, with each additional timestamp reinforcing the ones before it. Hash Hash Owner 2's Signature Owner 1's Private Key Owner 2's Private Key Owner 3's Private Key Block Item Item ... 2 Block Item Item ... Verify Verify Sign Sign 4. Proof-of-Work To implement a distributed timestamp server on a peer-to-peer basis, we will need to use a proof- of-work system similar to Adam Back's Hashcash [6], rather than newspaper or Usenet posts. The proof-of-work involves scanning for a value that when hashed, such as with SHA-256, the hash begins with a number of zero bits. The average work required is exponential in the number of zero bits required and can be verified by executing a single hash. For our timestamp network, we implement the proof-of-work by incrementing a nonce in the block until a value is found that gives the block's hash the required zero bits. Once the CPU effort has been expended to make it satisfy the proof-of-work, the block cannot be changed without redoing the work. As later blocks are chained after it, the work to change the block would include redoing all the blocks after it. The proof-of-work also solves the problem of determining representation in majority decision making. If the majority were based on one-IP-address-one-vote, it could be subverted by anyone able to allocate many IPs. Proof-of-work is essentially one-CPU-one-vote. The majority decision is represented by the longest chain, which has the greatest proof-of-work effort invested in it. If a majority of CPU power is controlled by honest nodes, the honest chain will grow the fastest and outpace any competing chains. To modify a past block, an attacker would have to redo the proof-of-work of the block and all blocks after it and then catch up with and surpass the work of the honest nodes. We will show later that the probability of a slower attacker catching up diminishes exponentially as subsequent blocks are added. To compensate for increasing hardware speed and varying interest in running nodes over time, the proof-of-work difficulty is determined by a moving average targeting an average number of blocks per hour. If they're generated too fast, the difficulty increases. 5. Network The steps to run the network are as follows: 1) New transactions are broadcast to all nodes. 2) Each node collects new transactions into a block. 3) Each node works on finding a difficult proof-of-work for its block. 4) When a node finds a proof-of-work, it broadcasts the block to all nodes. 5) Nodes accept the block only if all transactions in it are valid and not already spent. 6) Nodes express their acceptance of the block by working on creating the next block in the chain, using the hash of the accepted block as the previous hash. Nodes always consider the longest chain to be the correct one and will keep working on extending it. If two nodes broadcast different versions of the next block simultaneously, some nodes may receive one or the other first. In that case, they work on the first one they received, but save the other branch in case it becomes longer. The tie will be broken when the next proof- of-work is found and one branch becomes longer; the nodes that were working on the other branch will then switch to the longer one. 3 Block Nonce Tx Tx ... Block Nonce Tx Tx ... Prev Hash Prev Hash New transaction broadcasts do not necessarily need to reach all nodes. As long as they reach many nodes, they will get into a block before long. Block broadcasts are also tolerant of dropped messages. If a node does not receive a block, it will request it when it receives the next block and realizes it missed one. 6. Incentive By convention, the first transaction in a block is a special transaction that starts a new coin owned by the creator of the block. This adds an incentive for nodes to support the network, and provides a way to initially distribute coins into circulation, since there is no central authority to issue them. The steady addition of a constant of amount of new coins is analogous to gold miners expending resources to add gold to circulation. In our case, it is CPU time and electricity that is expended. The incentive can also be funded with transaction fees. If the output value of a transaction is less than its input value, the difference is a transaction fee that is added to the incentive value of the block containing the transaction. Once a predetermined number of coins have entered circulation, the incentive can transition entirely to transaction fees and be completely inflation free. The incentive may help encourage nodes to stay honest. If a greedy attacker is able to assemble more CPU power than all the honest nodes, he would have to choose between using it to defraud people by stealing back his payments, or using it to generate new coins. He ought to find it more profitable to play by the rules, such rules that favour him with more new coins than everyone else combined, than to undermine the system and the validity of his own wealth. 7. Reclaiming Disk Space Once the latest transaction in a coin is buried under enough blocks, the spent transactions before it can be discarded to save disk space. To facilitate this without breaking the block's hash, transactions are hashed in a Merkle Tree [7][2][5], with only the root included in the block's hash. Old blocks can then be compacted by stubbing off branches of the tree. The interior hashes do not need to be stored. Block Hash0 Hash1 Hash2 Hash3 Tx0 Tx1 Tx2 Tx3 Block Header (Block Hash) Prev Hash Nonce Root Hash Hash01 Hash23 Block Block Header (Block Hash) Prev Hash Nonce Root Hash Hash01 Hash23 Hash2 Hash3 Tx3 Transactions Hashed in a Merkle Tree After Pruning Tx0-2 from the Block A block header with no transactions would be about 80 bytes. If we suppose blocks are generated every 10 minutes, 80 bytes * 6 * 24 * 365 = 4.2MB per year. With computer systems typically selling with 2GB of RAM as of 2008, and Moore's Law predicting current growth of 1.2GB per year, storage should not be a problem even if the block headers must be kept in memory. 4 8. Simplified Payment Verification It is possible to verify payments without running a full network node. A user only needs to keep a copy of the block headers of the longest proof-of-work chain, which he can get by querying network nodes until he's convinced he has the longest chain, and obtain the Merkle branch linking the transaction to the block it's timestamped in. He can't check the transaction for himself, but by linking it to a place in the chain, he can see that a network node has accepted it, and blocks added after it further confirm the network has accepted it. Longest Proof-of-Work Chain Block Header Block Header Block Header Prev Hash Nonce Prev Hash Nonce Prev Hash Nonce Merkle Root Merkle Root Merkle Root Hash01 Hash23 Merkle Branch for Tx3 Hash2 Hash3 Tx3 As such, the verification is reliable as long as honest nodes control the network, but is more vulnerable if the network is overpowered by an attacker. While network nodes can verify transactions for themselves, the simplified method can be fooled by an attacker's fabricated transactions for as long as the attacker can continue to overpower the network. One strategy to protect against this would be to accept alerts from network nodes when they detect an invalid block, prompting the user's software to download the full block and alerted transactions to confirm the inconsistency. Businesses that receive frequent payments will probably still want to run their own nodes for more independent security and quicker verification. 9. Combining and Splitting Value Although it would be possible to handle coins individually, it would be unwieldy to make a separate transaction for every cent in a transfer. To allow value to be split and combined, transactions contain multiple inputs and outputs. Normally there will be either a single input from a larger previous transaction or multiple inputs combining smaller amounts, and at most two outputs: one for the payment, and one returning the change, if any, back to the sender. It should be noted that fan-out, where a transaction depends on several transactions, and those transactions depend on many more, is not a problem here. There is never the need to extract a complete standalone copy of a transaction's history. 5 Transaction In Out In ... ... 10. Privacy The traditional banking model achieves a level of privacy by limiting access to information to the parties involved and the trusted third party. The necessity to announce all transactions publicly precludes this method, but privacy can still be maintained by breaking the flow of information in another place: by keeping public keys anonymous. The public can see that someone is sending an amount to someone else, but without information linking the transaction to anyone. This is similar to the level of information released by stock exchanges, where the time and size of individual trades, the "tape", is made public, but without telling who the parties were. Traditional Privacy Model Identities Transactions New Privacy Model Identities Transactions As an additional firewall, a new key pair should be used for each transaction to keep them from being linked to a common owner. Some linking is still unavoidable with multi-input transactions, which necessarily reveal that their inputs were owned by the same owner. The risk is that if the owner of a key is revealed, linking could reveal other transactions that belonged to the same owner. 11. Calculations We consider the scenario of an attacker trying to generate an alternate chain faster than the honest chain. Even if this is accomplished, it does not throw the system open to arbitrary changes, such as creating value out of thin air or taking money that never belonged to the attacker. Nodes are not going to accept an invalid transaction as payment, and honest nodes will never accept a block containing them. An attacker can only try to change one of his own transactions to take back money he recently spent. The race between the honest chain and an attacker chain can be characterized as a Binomial Random Walk. The success event is the honest chain being extended by one block, increasing its lead by +1, and the failure event is the attacker's chain being extended by one block, reducing the gap by -1. The probability of an attacker catching up from a given deficit is analogous to a Gambler's Ruin problem. Suppose a gambler with unlimited credit starts at a deficit and plays potentially an infinite number of trials to try to reach breakeven. We can calculate the probability he ever reaches breakeven, or that an attacker ever catches up with the honest chain, as follows [8]: p = probability an honest node finds the next block q = probability the attacker finds the next block qz = probability the attacker will ever catch up from z blocks behind Trusted Third Party q ={ 1 if p≤q} z q/pz if pq 6 Counterparty Public Public Given our assumption that p > q, the probability drops exponentially as the number of blocks the attacker has to catch up with increases. With the odds against him, if he doesn't make a lucky lunge forward early on, his chances become vanishingly small as he falls further behind. We now consider how long the recipient of a new transaction needs to wait before being sufficiently certain the sender can't change the transaction. We assume the sender is an attacker who wants to make the recipient believe he paid him for a while, then switch it to pay back to himself after some time has passed. The receiver will be alerted when that happens, but the sender hopes it will be too late. The receiver generates a new key pair and gives the public key to the sender shortly before signing. This prevents the sender from preparing a chain of blocks ahead of time by working on it continuously until he is lucky enough to get far enough ahead, then executing the transaction at that moment. Once the transaction is sent, the dishonest sender starts working in secret on a parallel chain containing an alternate version of his transaction. The recipient waits until the transaction has been added to a block and z blocks have been linked after it. He doesn't know the exact amount of progress the attacker has made, but assuming the honest blocks took the average expected time per block, the attacker's potential progress will be a Poisson distribution with expected value: = z qp To get the probability the attacker could still catch up now, we multiply the Poisson density for each amount of progress he could have made by the probability he could catch up from that point: ∞ ke−{q/pz−k ifk≤z} ∑k=0 k!⋅ 1 ifkz Rearranging to avoid summing the infinite tail of the distribution... z ke− z−k 1−∑k=0 k! 1−q/p Converting to C code... #include <math.h> double AttackerSuccessProbability(double q, int z) { double p = 1.0 - q; double lambda = z * (q / p); double sum = 1.0; int i, k; for (k = 0; k <= z; k++) { double poisson = exp(-lambda); for (i = 1; i <= k; i++) poisson *= lambda / i; sum -= poisson * (1 - pow(q / p, z - k)); } return sum; } 7 Running some results, we can see the probability drop off exponentially with z. q=0.1 z=0 P=1.0000000 z=1 P=0.2045873 z=2 P=0.0509779 z=3 P=0.0131722 z=4 P=0.0034552 z=5 P=0.0009137 z=6 P=0.0002428 z=7 P=0.0000647 z=8 P=0.0000173 z=9 P=0.0000046 z=10 P=0.0000012 q=0.3 z=0 P=1.0000000 z=5 P=0.1773523 z=10 P=0.0416605 z=15 P=0.0101008 z=20 P=0.0024804 z=25 P=0.0006132 z=30 P=0.0001522 z=35 P=0.0000379 z=40 P=0.0000095 z=45 P=0.0000024 z=50 P=0.0000006 Solving for P less than 0.1%... P < 0.001 q=0.10 z=5 q=0.15 z=8 q=0.20 z=11 q=0.25 z=15 q=0.30 z=24 q=0.35 z=41 q=0.40 z=89 q=0.45 z=340 12. Conclusion We have proposed a system for electronic transactions without relying on trust. We started with the usual framework of coins made from digital signatures, which provides strong control of ownership, but is incomplete without a way to prevent double-spending. To solve this, we proposed a peer-to-peer network using proof-of-work to record a public history of transactions that quickly becomes computationally impractical for an attacker to change if honest nodes control a majority of CPU power. The network is robust in its unstructured simplicity. Nodes work all at once with little coordination. They do not need to be identified, since messages are not routed to any particular place and only need to be delivered on a best effort basis. Nodes can leave and rejoin the network at will, accepting the proof-of-work chain as proof of what happened while they were gone. They vote with their CPU power, expressing their acceptance of valid blocks by working on extending them and rejecting invalid blocks by refusing to work on them. Any needed rules and incentives can be enforced with this consensus mechanism. 8 References [1] W. Dai, "b-money," http://www.weidai.com/bmoney.txt, 1998. [2] H. Massias, X.S. Avila, and J.-J. Quisquater, "Design of a secure timestamping service with minimal trust requirements," In 20th Symposium on Information Theory in the Benelux, May 1999. [3] S. Haber, W.S. Stornetta, "How to time-stamp a digital document," In Journal of Cryptology, vol 3, no 2, pages 99-111, 1991. [4] D. Bayer, S. Haber, W.S. Stornetta, "Improving the efficiency and reliability of digital time-stamping," In Sequences II: Methods in Communication, Security and Computer Science, pages 329-334, 1993. [5] S. Haber, W.S. Stornetta, "Secure names for bit-strings," In Proceedings of the 4th ACM Conference on Computer and Communications Security, pages 28-35, April 1997. [6] A. Back, "Hashcash - a denial of service counter-measure," http://www.hashcash.org/papers/hashcash.pdf, 2002. [7] R.C. Merkle, "Protocols for public key cryptosystems," In Proc. 1980 Symposium on Security and Privacy, IEEE Computer Society, pages 122-133, April 1980. [8] W. Feller, "An introduction to probability theory and its applications," 1957. 9
Kechisgift
<!-- markdownlint-disable no-inline-html --> <!-- markdownlint-disable ol-prefix --> # GRAVITON (aka the TEAL Blackbox Toolkit): Program Reporting and Testing via Dry Runs <img width="345" alt="http://cds.cern.ch/record/2315186/files/scoap3-fulltext.pdf" src="https://user-images.githubusercontent.com/291133/160721859-21a3560a-0a82-4249-aa54-5ede4c60f8d2.png"> **NOTE: to get math formulas to render here using Chrome, add the [xhub extension](https://chrome.google.com/webstore/detail/xhub/anidddebgkllnnnnjfkmjcaallemhjee/related) and reload** **DISCLAIMER**: Graviton is subject to change and makes no backwards compatability guarantees. ## Blackbox Testing Howto ### What is TEAL Blackbox Testing? TEAL Blackbox Testing lets you gain confidence that your Algorand smart contracts are correct by writing assertions and and analyzing results via dry runs. ### Why Blackbox Testing? Here are some use cases: * by allowing you to assert that certain invariants hold over a large set of inputs you gain greater confidence that your TEAL programs and AVM smart contracts work as designed * when tweaking, refactoring or optimizing your TEAL source, ensure that no regressions have occured * allows AVM developers to practice the art of TTDD (TEAL Test Driven Development) ## Simple TEAL Blackbox Toolkit Example: Program for $`x^2`$ Consider this [TEAL program](./tests/teal/lsig_square.teal) for computing $`x^2`$: ```plain #pragma version 6 arg 0 btoi callsub square_0 return // square square_0: store 0 load 0 pushint 2 // 2 exp retsub ``` We'd like to write some unit tests to validate its correctness and make **assertions** about the: * program's opcode cost * program's stack * stack's height * scratch variables * final log message (this is especially useful for [ABI-compliant programs](https://developer.algorand.org/docs/get-details/dapps/smart-contracts/ABI/)) * status (**PASS** or **REJECT**) * error conditions that are and aren't encountered Even better, before making fine-grained assertions we'd like to get a sense of what the program is doing on a large set of inputs and discover _experimentally_ these **program invariants**. Let's go through how we can do this: * start by making basic assertions and validate them using dry runs (see "**Basic Assertions**" section below) * execute the program on a sequence of inputs and explore the results (see "**EDRA: Exploratory Dry Run Analysis**" section below) * create invariants for the entire sequence and assert that the invariants hold (see "**Advanced: Asserting Invariants on a Dry Run Sequence**" section below) > Becoming a TEAL Blackbox Toolkit Ninja involves 10 steps as described below ### Dry Run Environment Setup **STEP 1**. Start with a running local node and make note of Algod's port number (for our [standard sandbox](https://github.com/algorand/sandbox) this is `4001`) **STEP 2**. Set the `ALGOD_PORT` value in [tests/clients.py](./tests/clients.py#L7) to this port number. (The port is already pre-set to `4001` because [graviton](https://github.com/algorand/graviton)'s [CI process](https://en.wikipedia.org/wiki/Continuous_integration) uses the standad sandbox) ### TEAL Program for Testing: Logic Sig v. App **STEP 3**. Next, you'll need to figure out if your TEAL program should be a Logic Signature or an Application. Each of these program _modes_ has its merits, but we won't get into the pros/cons here. From a Blackbox Test's perspective, the main difference is how external arguments are handled. Logic sigs rely on the [arg opcode](https://developer.algorand.org/docs/get-details/dapps/avm/teal/opcodes/#arg-n) while apps rely on [txna ApplicationArgs i](https://developer.algorand.org/docs/get-details/dapps/avm/teal/opcodes/#txna-f-i). In our $`x^2`$ **logic sig** example, you can see on [line 2](./tests/teal/lsig_square.teal#L2) that the `arg` opcode is used. Because each argument opcode (`arg` versus `ApplicationArgs`) is mode-exclusive, any program that takes input will execute succesfully in _one mode only_. **STEP 4**. Write the TEAL program that you want to test. You can inline the test as described here or follow the approach of `./tests/integration/blackbox_test.py` and save under `./tests/teal`. So following the inline appraoch we begin our TEAL Blackbox script with an <a name="teal">inline teal source variable</a>: ```python teal = """#pragma version 6 arg 0 btoi callsub square_0 return // square square_0: store 0 load 0 pushint 2 // 2 exp retsub""" ``` ### The TEAL Blackbox Toolkit's Utitlity Classes The TEAL Blackbox Toolkit comes with the following main classes: * `DryRunExecutor` - executes dry run's for apps and logic sigs for one or more inputs * `DryRunInspector` - encapsulates a dry run's result for a single input and allows inspecting and making assertions about it * `Invariant` - class for asserting invariants about a _sequence_ of dry run executions in a declarative fashion ### Basic Assertions When executing a dry run using `DryRunExecutor` you'll get back `DryRunInspector` objects. Such objects have **assertable properties** which can be used to validate the dry run. **STEP 4**. Back to our $`x^2`$ example, and assuming the `teal` variable is defined [as above](#teal). You can run the following: ```python from graviton.blackbox import DryRunExecutor from tests.clients import get_algod algod = get_algod() x = 9 args = (x,) inspector = DryRunExecutor.dryrun_logicsig(algod, teal, args) assert inspector.status() == "PASS" assert inspector.stack_top() == x**2 ``` Here we have executed a dry run on input $`x=9`$, then asserted that: * the program status was `PASS` * the program exited with the top of its stack containing $`x^2 = 9^2 = 81`$ Some available _assertable properties_ are: * `stack_top()` * `last_log()` * `cost()` * `status()` * `final_scratch()` * `error()` * `max_stack_height()` See the [DryRunInspector class comment](./graviton/blackbox.py#L387) for more assertable properties and details. ### Printing out the TEAL Stack Trace for a Failing Assertion **STEP 5**. The `DryRunInspector`'s `report()` method lets you print out a handy report in the case of a failing assertion. Let's intentionally break the test case above by claiming that $`x^2 = x^3`$ for $`x=2`$ and print out this _report_ when our silly assertion fails. ```python from graviton.blackbox import DryRunExecutor from tests.clients import get_algod algod = get_algod() x = 2 args = (x,) inspector = DryRunExecutor.dryrun_logicsig(algod, teal, args) # This one's ok expected, actual = "PASS", inspector.status() assert expected == actual, inspector.report( args, f"expected {expected} but got {actual}" ) # This one's absurd! x^3 != x^2 expected, actual = x**3, inspector.stack_top() assert expected == actual, inspector.report( args, f"expected {expected} but got {actual}" ) ``` If we run the test (e.g. with `pytest`) we'll see a printout such as: ```sh AssertionError: =============== <<<<<<<<<<<expected 8 but got 4>>>>>>>>>>> =============== App Trace: step | PC# | L# | Teal | Scratch | Stack --------+-------+------+-------------------+-----------+---------------------- 1 | 1 | 1 | #pragma version 6 | | [] 2 | 2 | 2 | arg_0 | | [0x0000000000000002] 3 | 3 | 3 | btoi | | [2] 4 | 7 | 6 | label1: | | [2] 5 | 9 | 7 | store 0 | 0->2 | [] 6 | 11 | 8 | load 0 | | [2] 7 | 13 | 9 | pushint 2 | | [2, 2] 8 | 14 | 10 | exp | | [4] 9 | 6 | 4 | callsub label1 | | [4] 10 | 15 | 11 | retsub | | [4] =============== MODE: ExecutionMode.Signature TOTAL COST: None =============== FINAL MESSAGE: PASS =============== Messages: ['PASS'] Logs: [] =============== -----BlackBoxResult(steps_executed=10)----- TOTAL STEPS: 10 FINAL STACK: [4] FINAL STACK TOP: 4 MAX STACK HEIGHT: 2 FINAL SCRATCH: {0: 2} SLOTS USED: [0] FINAL AS ROW: {'steps': 10, ' top_of_stack': 4, 'max_stack_height': 2, 's@000': 2} =============== Global Delta: [] =============== Local Delta: [] =============== TXN AS ROW: {' Run': 0, ' cost': None, ' last_log': '`None', ' final_message': 'PASS', ' Status': 'PASS', 'steps': 10, ' top_of_stack': 4, 'max_stack_height': 2, 's@000': 2, 'Arg_00': 2} =============== <<<<<<<<<<<expected 8 but got 4>>>>>>>>>>> =============== assert 8 == 4 ``` In particular, we can: * Track the program execution by viewing its **App Trace** * 2 was assigned to **scratch slot #0** at step 5 * the stack ended up with **4** on top * the run **PASS**'ed * Read the message parameter that was provided and which explains in English what went wrong: `expected 8 but got 4` ### EDRA: Exploratory Dry Run Analysis Let's expand our investigation from a single dry-run to multiple runs or a **run sequence**. We'll observe how _assertable properties_ depend on inputs and conjecture some program invariants. To aid in the investigation we'll generate a report in CSV format (Comma Separated Values) where: * columns represent _assertable properties_ of dry-runs, and * rows represents dry-run executions for specific inputs **STEP 6**. Back to our $`x^2`$ example, here's how to generate a report with 1 row for each of the inputs `0, 1, ... , 15`: ```python from graviton.blackbox import DryRunExecutor, DryRunInspector from tests.clients import get_algod algod = get_algod() inputs = [(x,) for x in range(16)] run_results = DryRunExecutor.dryrun_logicsig_on_sequence(algod, teal, inputs) csv = DryRunInspector.csv_report(inputs, run_results) print(csv) ``` Note that each element in the `inputs` array `(x,)` is itself a tuple as `args` given to a dry run execution need to be of type `Sequence` (remember, that these will be passed to a TEAL program which may take one, several, or no inputs at all). At this point, you'll be able to look at your [dry run sequence results](./graviton/blackbox.py#L752) and conduct some analysis. For the $`x^2`$ example, after loading the CSV in Google sheets and reformating a bit it will look like: <img width="465" alt="image" src="https://user-images.githubusercontent.com/291133/158812699-318169e2-487c-4dac-b97b-a9db8148b638.png"> Pointing out some interesting results: * column `D` **Arg 00** has the input $`x`$ (it's the argument at index 0) * column `A` contains the **Run** number * column `E` **top of stack** is the value at program's termination, i.e. $`x^2`$ * column `B` **Status** of each runs **PASS**es _except for **Run 1** with **Arg 00** = 0_. (The first run **REJECT**s because $`0^2 = 0`$ and TEAL programs reject when the top of the stack is 0) * column `G` shows scratch slot **s@000** which stores the value of $`x`$ (except for the case $`x = 0`$ in which appears empty; in fact, slots always default to the zero value and an **<a name="0val-artifact">artifact</a>** of dry-runs is that they do not report when 0-values get stored into previously empty slots as no state change actually occurs) * column `F` **max stack height** is always 2. The final observation makes sense because there is no branching or looping in the program. **STEP 7**. We can re-cast the observed effects in `Columns E, B, G, F` as **invariants** written in Python as follows: * `inspector.stack_top() == x ** 2` * `inspector.max_stack_height() == 2` * `inspector.status() == ("REJECT" if x == 0 else "PASS")` * `inspector.final_scratch() == ({} if x == 0 else {0: x})` ### Advanced: Asserting Invariants on a Dry Run Sequence The final and most advanced topic we'll cover is how to assert that invariants hold on a sequence of inputs. Lets take the information we gleaned in our EDRA CSV report, and create an integration test out of it. There are two ways to achieve this goal: * Procedural invariant assertions * Declarative invariant assertions #### Procedural Blackbox Dry Run Sequence Assertions **STEP 8**. The procedural approach takes the _invariants_ and simply asserts them inside of a for loop that iterates over the inputs and dry runs. One can call each dry run execution independently, or use `DryRunExecutor`'s convenience methods `dryrun_app_on_sequence()` and `dryrun_logicsig_on_sequence()`. For example, let's assert that the above invariants hold for all $`x \leq 100`$: ```python from graviton.blackbox import DryRunExecutor from tests.clients import get_algod algod = get_algod() inputs = [(x,) for x in range(101)] dryrun_results = DryRunExecutor.dryrun_logicsig_on_sequence(algod, teal, inputs) for i, inspector in enumerate(dryrun_results): args = inputs[i] x = args[0] assert inspector.stack_top() == x**2 assert inspector.max_stack_height() == 2 assert inspector.status() == ("REJECT" if x == 0 else "PASS") assert inspector.final_scratch() == ({} if x == 0 else {0: x}) ``` #### Declarative Blackbox Dry Run Sequence Assertions **STEP 9**. The TEAL Blackbox Toolkit also allows for declarative style test writing. Let's define some invariants for a particular sequence of `lsig_square` TEAL program dry runs: ```python scenario = { "inputs": [(i,) for i in range(100)], "invariants": { DRProp.stackTop: lambda args: args[0] ** 2, DRProp.maxStackHeight: 2, DRProp.status: lambda i: "REJECT" if i[0] = 0 else "PASS", DRProp.finalScratch: lambda args: ({} if args[0] else {0: args[0]}), }, } ``` In the parlance of the TEAL Blackbox Toolkit, a set of such declarative assertions is called a **test scenario**. Scenarios are dict's containing two keys `inputs` and `invariants` and follow [certain conventions](./graviton/invariant.py#L101). In particular: * **inputs** gives a list of tuples, each tuple representing the `args` to be fed into a single dry run execution * **invariants** gives a dict that maps [DryRunProperty](./graviton/blackbox.py#L25)'s to an invariant _predicate_ In English, letting $`x`$ be the input variable for our square function, the above **test scenario**: * provides a list of 100 tuples of the form $`(x)`$ that will serve as args. * IE: $`(0), (1), (2), ... , (99)`$ * establishes 4 different _invariants_ as follows: * the **stack's top** will contain $`x^2`$ * the **max stack height** during execution is always 2 * the executions' **status** is **PASS** except for the case $`x=0`$ * the **final scratch** will have $`x`$ stored at slot `0` except for that strange $`x=0`$ case (recall the [0-val scratch slot artifact](#0val-artifact)) Declarative invariants make use of the following: * `DryRunProperty` (aka `DRProp`): an enum that acts as a key in a scenario's assertions dict * class `Invariant` * its constructor takes in a predicate (there are [4 kinds of predicates](#predicate)) and returns a callable that is used for runtime assertions * method `inputs_and_assertions()` validates a scenario and extracts out its assertions * method `dryrun_assert()` evaluates the dry-run sequence using the constructed `SequenceAssertion` To employ the declarative test scenario above write the following: ```python from graviton.blackbox import ( DryRunExecutor, DryRunProperty as DRProp, ExecutionMode, ) from graviton.invariant import Invariant from tests.clients import get_algod algod = get_algod() scenario = { "inputs": [(i,) for i in range(100)], "invariants": { DRProp.stackTop: lambda args: args[0] ** 2, DRProp.maxStackHeight: 2, DRProp.status: lambda args: "REJECT" if args[0] == 0 else "PASS", DRProp.finalScratch: lambda args: ({0: args[0]} if args[0] else {}), }, } # Validate the scenario and dig out inputs/invariants: inputs, invariants = Invariant.inputs_and_invariants( scenario, ExecutionMode.Signature ) # Execute the dry runs and obtain a sequence of DryRunInspectors: inspectors = DryRunExecutor.dryrun_logicsig_on_sequence(algod, teal, inputs) # Invariant assertions on sequence: for dr_property, invariant in invariants.items(): invariant.validates(dr_property, inputs, inspectors) ``` **STEP 10**. _**Deep Dive into Invariants via Exercises**_ Four kinds of <a name="predicate">predicates</a> are used to define _invariants_: 1. _simple python types_ - these are useful in the case of _constant_ invariants. In the above `maxStackHeight` is asserted to _**ALWAYS**_ equal 2 by using `2` in the declaration: `DRProp.maxStackHeight: 2` 2. _1-variable functions_ -these are useful when you have a python "simulator" for the invariant. In the above `stackTop` is asserted to be $`x^2`$ by using a lambda expression for $`x^2`$ in the declaration: `DRProp.stackTop: lambda args: args[0] ** 2` 3. _dictionaries_ of type `Dict[Tuple, Any]` - these are useful when you want to assert a discrete set of input-output pairs. For example, if you have 4 inputs that you want to assert are being squared, you could use ```python DRProp.stackTop: { (2,): 4, (7,): 49, (13,): 169, (11,): 121, } ``` >Note that this case illustrates why each `args` container should be a tuple intead of a list. In order to specify a map from args to expected, we need to make `args` a key in a dictionary. Python dictionary keys must be hashable and lists are _not hashable_ while tuples _are_ hashable, hence the tuple-requirement. 4. _2-variable functions_ -these are useful when your assertion is more subtle than out-and-out equality. For example, suppose you want to assert that the `cost` of each run is _between_ $`2n \pm 5`$ where $`n`$ is the first arg of the input. Then you could declare: `DRProp.cost: lambda args, actual: 2*args[0] - 5 <= actual <= 2*args[0] + 5` #### **EXERCISE A** Convert each of the lambda expressions used above to dictionaries that assert the same thing. #### **EXERCISE B** Use 2-variable functions in order to _ignore_ the weird $`x=0`$ cases above. #### _PARTIAL SOLUTIONS to EXERCISES_ **Exercise A Partial Solution**. For `DRProp.status`'s declaration you could define the `dict` using dictionary comprehension syntax as follows: ```python DRProp.status: {(x,): "PASS" if x else "REJECT" for x in range(100)}, ``` **Exercise B Partial Solution**. For `DRProp.status`'s declaration you could ignore the case $`x=0`$ as follows: ```python DRProp.status: lambda args, actual: "PASS" == actual if args[0] else True, ``` ## Slow and Bad Fibonacci - Another Example Report [This](https://docs.google.com/spreadsheets/d/1ax-jQdYCkKT61Z6SPeGm5BqAMybgkWJa-Dv0yVjgFSA/edit?usp=sharing) is an example of `app_slow_fibonacci.teal`'s Dryrun stats: <img width="1231" alt="image" src="https://user-images.githubusercontent.com/291133/158705149-302d755f-afcc-4380-976a-ca14800c138f.png"> A few items to take note of: * $`n`$ is given by **Arg_00** * the app was **REJECT**ed for $`n = 0`$ because `fibonacci(0) == 0` is left at the top of the stack * the app was **REJECT**ed for $`n > 7`$ because of exceeding budget * the app **errored** only for $`n > 16`$ because of exceeding _dynamic_ budget * the **cost** is growing exponentially (poor algorithm design) * the **top of stack** contains `fibonacci(n)` except in the error case * the **final_log** contains `hex(fibonacci(n))` except in the error and reject cases * **max stack height** is $`2n`$ except for $`n=0`$ and the error case * you can see the final values of scratch slots **s@000** and **s@001** which are respectively $`n`$ and `fibonacci(n)` Here's an example of how [invariants can be asserted](https://github.com/algorand/graviton/blob/a8c7eab729a36503948849674ea55995d5fc4ec1/tests/integration/blackbox_test.py#L315) on this function.
umar-briamaah
A powerful Python tool for extracting password hashes from encrypted PDF files for use with password cracking tools like Hashcat and John the Ripper.
incendiary
PDF2John extracts a hash from an encrypted PDF for cracking. This is a dockerised version containing the libraries used.
joaohmenezes
C++ tool that parses entire directories of PDF files to extract text, count total words per file, and calculate alphabetical frequency using Hash Maps.
maximumgravity1
Proof-of-concept for a Universal PDF Container (UPC) — pack any binary file (like RAW images) inside a PDF, then extract and verify it with SHA256 hashes. Includes demo scripts for end-to-end workflow: create → embed → recover → validate.
MohamadAminSharifi
A general-purpose smart web crawler that recursively extracts website content using BFS, preserves full URL-based directory structure, and saves files locally. Supports PDF/Word/Excel/PowerPoint also media extraction, customizable crawl depth, domain restriction, CSS-based content targeting, and duplicate prevention using hashing.
thelkotolsantosh
Purpose: Core Python application for PDF hash extraction Key Features: - Extract hashes from single or multiple PDFs - Integration with John the Ripper - Hashcat compatibility verification - JSON report generation - Comprehensive error handling
easonlai
Extract PDF Text Content and Perform Hashing
BillReyor
Hash Hound is a tool for extracting hashes from various text and PDF reports.
xtraeme-xt
Random scripts to extract PDF security details and manipulate hashes
Donny-GUI
script for automatically extracting pdf file hashes for running hashcat on
ali-raheem
Extract the hash from an encrypted PDF in a format suitable for John the Ripper or Hashcat
Artanomell
Extract observables (hashes, ips, urls, etc) from various input formats (.pdf, .doc, .docx, .xls, .xlsx, .txt, .csv, etc)
mateusKing123
Forensic Extractor é uma API simples de análise forense digital que permite extrair metadados de arquivos, calacular hashes e gerar relatórios JSON ou PDF.
Lx050
Intelligent document image asset extraction, classification, and deduplication toolkit. Extract embedded images from DOCX/PDF, auto-classify with VLM, deduplicate with perceptual hashing.
yairoman
FastAPI service for extracting text and structured content from PDFs and images, with OCR, Gemini AI support, MongoDB persistence, hash-based caching, and automatic file retention cleanup.
davidhc1230
Video Page Extractor is a tool for extracting stable frames from videos, filtering out duplicates using Perceptual Hashing (PHash) and Structural Similarity Index (SSIM). It efficiently selects key frames while reducing redundant images, making it ideal for scanning PDFs, analyzing presentations, or capturing animation frames.
NANDHAKUMAR-ARJUNAN
ZeroTrace Certificate Verification Tool A lightweight, offline utility to verify cryptographic erasure certificates. It extracts the embedded JSON, hash, and digital signature from a signed PDF, recomputes the SHA-256 hash, and validates authenticity using the official public key.
nobody-Justheader
Hashcatizer — Single-binary Rust tool that extracts hashes from 89 encrypted file formats (SSH keys, KeePass, PDF, Office, crypto wallets, disk images, and more) and outputs them ready to crack with hashcat.
Matt1Up
Drop pdf files from the "MCRO Evidentiary Dataset" into the folder and all internal objects are extracted, hashed, and inserted into a csv table along with digital signature data (pdfsig) and metadata author and creator (exiftool)
akashkannancybersec
This project is a **Flask web application** that allows users to **upload files** and extract **detailed metadata** about them. It supports images, videos, documents (PDF/DOCX), and MP3 files. Metadata includes **file details, hashes, EXIF, GPS, media info, and more**.
tejasbargujepatil
FileFort is a powerful GUI-based tool for extracting and enriching Indicators of Compromise (IOCs) like hashes, IPs, URLs, and domains from files (PDF, JSON, CSV, etc.). It integrates with VirusTotal for threat enrichment and supports exporting results to CSV/JSON. Secure your files with ease!
BalaSravya1404
Comprehensive Automated Document Verification is a Flask-based system that securely generates and verifies academic documents. It hashes structured student data and profile images using SHA-256, embeds them in a QR-coded PDF, and validates integrity by re-extracting text, images, and QR data to detect tampering or unauthorized modifications.
All 25 repositories loaded