Diff of the two buildlogs: -- --- b1/build.log 2021-07-17 19:24:10.962407430 +0000 +++ b2/build.log 2021-07-17 19:49:20.745588141 +0000 @@ -1,6 +1,6 @@ I: pbuilder: network access will be disabled during build -I: Current time: Fri Aug 19 13:20:47 -12 2022 -I: pbuilder-time-stamp: 1660958447 +I: Current time: Sun Jul 18 09:24:18 +14 2021 +I: pbuilder-time-stamp: 1626549858 I: Building the build Environment I: extracting base tarball [/var/cache/pbuilder/bullseye-reproducible-base.tgz] I: copying local configuration @@ -16,8 +16,8 @@ I: copying [./caffe_1.0.0+git20180821.99bd997-8.debian.tar.xz] I: Extracting source gpgv: unknown type of key resource 'trustedkeys.kbx' -gpgv: keyblock resource '/tmp/dpkg-verify-sig.GBoGgFwK/trustedkeys.kbx': General error -gpgv: Signature made Fri Nov 20 19:51:48 2020 -12 +gpgv: keyblock resource '/tmp/dpkg-verify-sig.6eVLd3Di/trustedkeys.kbx': General error +gpgv: Signature made Sat Nov 21 21:51:48 2020 +14 gpgv: using RSA key 638BC75EC1E5C589067E35DE62645EB35F686A8A gpgv: issuer "lumin@debian.org" gpgv: Can't check signature: No public key @@ -34,135 +34,169 @@ dpkg-source: info: applying pyrequire-no-leveldb.patch I: using fakeroot in build. I: Installing the build-deps -I: user script /srv/workspace/pbuilder/29121/tmp/hooks/D02_print_environment starting +I: user script /srv/workspace/pbuilder/3330/tmp/hooks/D01_modify_environment starting +debug: Running on codethink16-arm64. +I: Changing host+domainname to test build reproducibility +I: Adding a custom variable just for the fun of it... +I: Changing /bin/sh to bash +Removing 'diversion of /bin/sh to /bin/sh.distrib by dash' +Adding 'diversion of /bin/sh to /bin/sh.distrib by bash' +Removing 'diversion of /usr/share/man/man1/sh.1.gz to /usr/share/man/man1/sh.distrib.1.gz by dash' +Adding 'diversion of /usr/share/man/man1/sh.1.gz to /usr/share/man/man1/sh.distrib.1.gz by bash' +I: Setting pbuilder2's login shell to /bin/bash +I: Setting pbuilder2's GECOS to second user,second room,second work-phone,second home-phone,second other +I: user script /srv/workspace/pbuilder/3330/tmp/hooks/D01_modify_environment finished +I: user script /srv/workspace/pbuilder/3330/tmp/hooks/D02_print_environment starting I: set - BUILDDIR='/build' - BUILDUSERGECOS='first user,first room,first work-phone,first home-phone,first other' - BUILDUSERNAME='pbuilder1' - BUILD_ARCH='arm64' - DEBIAN_FRONTEND='noninteractive' + BASH=/bin/sh + BASHOPTS=checkwinsize:cmdhist:complete_fullquote:extquote:force_fignore:globasciiranges:hostcomplete:interactive_comments:progcomp:promptvars:sourcepath + BASH_ALIASES=() + BASH_ARGC=() + BASH_ARGV=() + BASH_CMDS=() + BASH_LINENO=([0]="12" [1]="0") + BASH_SOURCE=([0]="/tmp/hooks/D02_print_environment" [1]="/tmp/hooks/D02_print_environment") + BASH_VERSINFO=([0]="5" [1]="1" [2]="4" [3]="1" [4]="release" [5]="aarch64-unknown-linux-gnu") + BASH_VERSION='5.1.4(1)-release' + BUILDDIR=/build + BUILDUSERGECOS='second user,second room,second work-phone,second home-phone,second other' + BUILDUSERNAME=pbuilder2 + BUILD_ARCH=arm64 + DEBIAN_FRONTEND=noninteractive DEB_BUILD_OPTIONS='buildinfo=+all reproducible=+all,-fixfilepath parallel=8' - DISTRIBUTION='' - HOME='/var/lib/jenkins' - HOST_ARCH='arm64' + DIRSTACK=() + DISTRIBUTION= + EUID=0 + FUNCNAME=([0]="Echo" [1]="main") + GROUPS=() + HOME=/var/lib/jenkins + HOSTNAME=i-capture-the-hostname + HOSTTYPE=aarch64 + HOST_ARCH=arm64 IFS=' ' - LANG='C' - LANGUAGE='en_US:en' - LC_ALL='C' - MAIL='/var/mail/root' - OPTIND='1' - PATH='/usr/sbin:/usr/bin:/sbin:/bin:/usr/games' - PBCURRENTCOMMANDLINEOPERATION='build' - PBUILDER_OPERATION='build' - PBUILDER_PKGDATADIR='/usr/share/pbuilder' - PBUILDER_PKGLIBDIR='/usr/lib/pbuilder' - PBUILDER_SYSCONFDIR='/etc' - PPID='29121' - PS1='# ' - PS2='> ' + LANG=C + LANGUAGE=nl_BE:nl + LC_ALL=C + MACHTYPE=aarch64-unknown-linux-gnu + MAIL=/var/mail/root + OPTERR=1 + OPTIND=1 + OSTYPE=linux-gnu + PATH=/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/i/capture/the/path + PBCURRENTCOMMANDLINEOPERATION=build + PBUILDER_OPERATION=build + PBUILDER_PKGDATADIR=/usr/share/pbuilder + PBUILDER_PKGLIBDIR=/usr/lib/pbuilder + PBUILDER_SYSCONFDIR=/etc + PIPESTATUS=([0]="0") + POSIXLY_CORRECT=y + PPID=3330 PS4='+ ' - PWD='/' - SHELL='/bin/bash' - SHLVL='2' - SUDO_COMMAND='/usr/bin/timeout -k 18.1h 18h /usr/bin/ionice -c 3 /usr/bin/nice /usr/sbin/pbuilder --build --configfile /srv/reproducible-results/rbuild-debian/tmp.4frKnk2Lej/pbuilderrc_QqoN --hookdir /etc/pbuilder/first-build-hooks --debbuildopts -b --basetgz /var/cache/pbuilder/bullseye-reproducible-base.tgz --buildresult /srv/reproducible-results/rbuild-debian/tmp.4frKnk2Lej/b1 --logfile b1/build.log caffe_1.0.0+git20180821.99bd997-8.dsc' - SUDO_GID='117' - SUDO_UID='110' - SUDO_USER='jenkins' - TERM='unknown' - TZ='/usr/share/zoneinfo/Etc/GMT+12' - USER='root' - USERNAME='root' - _='/usr/bin/systemd-run' - http_proxy='http://192.168.101.16:3128' + PWD=/ + SHELL=/bin/bash + SHELLOPTS=braceexpand:errexit:hashall:interactive-comments:posix + SHLVL=3 + SUDO_COMMAND='/usr/bin/timeout -k 24.1h 24h /usr/bin/ionice -c 3 /usr/bin/nice -n 11 /usr/bin/unshare --uts -- /usr/sbin/pbuilder --build --configfile /srv/reproducible-results/rbuild-debian/tmp.4frKnk2Lej/pbuilderrc_16oW --hookdir /etc/pbuilder/rebuild-hooks --debbuildopts -b --basetgz /var/cache/pbuilder/bullseye-reproducible-base.tgz --buildresult /srv/reproducible-results/rbuild-debian/tmp.4frKnk2Lej/b2 --logfile b2/build.log --extrapackages usrmerge caffe_1.0.0+git20180821.99bd997-8.dsc' + SUDO_GID=117 + SUDO_UID=110 + SUDO_USER=jenkins + TERM=unknown + TZ=/usr/share/zoneinfo/Etc/GMT-14 + UID=0 + USER=root + USERNAME=root + _='I: set' + http_proxy=http://192.168.101.16:3128 I: uname -a - Linux codethink13-arm64 4.15.0-147-generic #151-Ubuntu SMP Fri Jun 18 19:18:37 UTC 2021 aarch64 GNU/Linux + Linux i-capture-the-hostname 4.15.0-147-generic #151-Ubuntu SMP Fri Jun 18 19:18:37 UTC 2021 aarch64 GNU/Linux I: ls -l /bin total 5252 - -rwxr-xr-x 1 root root 1282512 Jun 21 2021 bash - -rwxr-xr-x 3 root root 34808 Jul 20 2020 bunzip2 - -rwxr-xr-x 3 root root 34808 Jul 20 2020 bzcat - lrwxrwxrwx 1 root root 6 Jul 20 2020 bzcmp -> bzdiff - -rwxr-xr-x 1 root root 2225 Jul 20 2020 bzdiff - lrwxrwxrwx 1 root root 6 Jul 20 2020 bzegrep -> bzgrep - -rwxr-xr-x 1 root root 4877 Sep 4 2019 bzexe - lrwxrwxrwx 1 root root 6 Jul 20 2020 bzfgrep -> bzgrep - -rwxr-xr-x 1 root root 3775 Jul 20 2020 bzgrep - -rwxr-xr-x 3 root root 34808 Jul 20 2020 bzip2 - -rwxr-xr-x 1 root root 14264 Jul 20 2020 bzip2recover - lrwxrwxrwx 1 root root 6 Jul 20 2020 bzless -> bzmore - -rwxr-xr-x 1 root root 1297 Jul 20 2020 bzmore - -rwxr-xr-x 1 root root 39832 Sep 22 2020 cat - -rwxr-xr-x 1 root root 64512 Sep 22 2020 chgrp - -rwxr-xr-x 1 root root 60368 Sep 22 2020 chmod - -rwxr-xr-x 1 root root 64528 Sep 22 2020 chown - -rwxr-xr-x 1 root root 138896 Sep 22 2020 cp - -rwxr-xr-x 1 root root 129544 Dec 10 2020 dash - -rwxr-xr-x 1 root root 101384 Sep 22 2020 date - -rwxr-xr-x 1 root root 80984 Sep 22 2020 dd - -rwxr-xr-x 1 root root 89824 Sep 22 2020 df - -rwxr-xr-x 1 root root 143088 Sep 22 2020 dir - -rwxr-xr-x 1 root root 76152 Feb 7 2021 dmesg - lrwxrwxrwx 1 root root 8 Nov 6 2019 dnsdomainname -> hostname - lrwxrwxrwx 1 root root 8 Nov 6 2019 domainname -> hostname - -rwxr-xr-x 1 root root 35632 Sep 22 2020 echo - -rwxr-xr-x 1 root root 28 Nov 9 2020 egrep - -rwxr-xr-x 1 root root 31512 Sep 22 2020 false - -rwxr-xr-x 1 root root 28 Nov 9 2020 fgrep - -rwxr-xr-x 1 root root 64856 Feb 7 2021 findmnt - -rwsr-xr-x 1 root root 34824 Feb 26 2021 fusermount - -rwxr-xr-x 1 root root 178400 Nov 9 2020 grep - -rwxr-xr-x 2 root root 2346 Mar 2 2021 gunzip - -rwxr-xr-x 1 root root 6376 Mar 2 2021 gzexe - -rwxr-xr-x 1 root root 93744 Mar 2 2021 gzip - -rwxr-xr-x 1 root root 18440 Nov 6 2019 hostname - -rwxr-xr-x 1 root root 68720 Sep 22 2020 ln - -rwxr-xr-x 1 root root 52720 Feb 7 2020 login - -rwxr-xr-x 1 root root 143088 Sep 22 2020 ls - -rwxr-xr-x 1 root root 161960 Feb 7 2021 lsblk - -rwxr-xr-x 1 root root 85200 Sep 22 2020 mkdir - -rwxr-xr-x 1 root root 68744 Sep 22 2020 mknod - -rwxr-xr-x 1 root root 43976 Sep 22 2020 mktemp - -rwxr-xr-x 1 root root 51368 Feb 7 2021 more - -rwsr-xr-x 1 root root 51360 Feb 7 2021 mount - -rwxr-xr-x 1 root root 14496 Feb 7 2021 mountpoint - -rwxr-xr-x 1 root root 134808 Sep 22 2020 mv - lrwxrwxrwx 1 root root 8 Nov 6 2019 nisdomainname -> hostname - lrwxrwxrwx 1 root root 14 Apr 18 2021 pidof -> /sbin/killall5 - -rwxr-xr-x 1 root root 35720 Sep 22 2020 pwd - lrwxrwxrwx 1 root root 4 Jun 21 2021 rbash -> bash - -rwxr-xr-x 1 root root 43872 Sep 22 2020 readlink - -rwxr-xr-x 1 root root 68592 Sep 22 2020 rm - -rwxr-xr-x 1 root root 43880 Sep 22 2020 rmdir - -rwxr-xr-x 1 root root 19208 Sep 27 2020 run-parts - -rwxr-xr-x 1 root root 114016 Dec 22 2018 sed - lrwxrwxrwx 1 root root 4 Aug 19 03:48 sh -> dash - -rwxr-xr-x 1 root root 35656 Sep 22 2020 sleep - -rwxr-xr-x 1 root root 72640 Sep 22 2020 stty - -rwsr-xr-x 1 root root 67776 Feb 7 2021 su - -rwxr-xr-x 1 root root 35672 Sep 22 2020 sync - -rwxr-xr-x 1 root root 535768 Feb 16 2021 tar - -rwxr-xr-x 1 root root 10568 Sep 27 2020 tempfile - -rwxr-xr-x 1 root root 89120 Sep 22 2020 touch - -rwxr-xr-x 1 root root 31512 Sep 22 2020 true - -rwxr-xr-x 1 root root 14264 Feb 26 2021 ulockmgr_server - -rwsr-xr-x 1 root root 30880 Feb 7 2021 umount - -rwxr-xr-x 1 root root 35640 Sep 22 2020 uname - -rwxr-xr-x 2 root root 2346 Mar 2 2021 uncompress - -rwxr-xr-x 1 root root 143088 Sep 22 2020 vdir - -rwxr-xr-x 1 root root 59584 Feb 7 2021 wdctl - lrwxrwxrwx 1 root root 8 Nov 6 2019 ypdomainname -> hostname - -rwxr-xr-x 1 root root 1984 Mar 2 2021 zcat - -rwxr-xr-x 1 root root 1678 Mar 2 2021 zcmp - -rwxr-xr-x 1 root root 5880 Mar 2 2021 zdiff - -rwxr-xr-x 1 root root 29 Mar 2 2021 zegrep - -rwxr-xr-x 1 root root 29 Mar 2 2021 zfgrep - -rwxr-xr-x 1 root root 2081 Mar 2 2021 zforce - -rwxr-xr-x 1 root root 7585 Mar 2 2021 zgrep - -rwxr-xr-x 1 root root 2206 Mar 2 2021 zless - -rwxr-xr-x 1 root root 1842 Mar 2 2021 zmore - -rwxr-xr-x 1 root root 4553 Mar 2 2021 znew -I: user script /srv/workspace/pbuilder/29121/tmp/hooks/D02_print_environment finished + -rwxr-xr-x 1 root root 1282512 Jun 22 16:26 bash + -rwxr-xr-x 3 root root 34808 Jul 21 2020 bunzip2 + -rwxr-xr-x 3 root root 34808 Jul 21 2020 bzcat + lrwxrwxrwx 1 root root 6 Jul 21 2020 bzcmp -> bzdiff + -rwxr-xr-x 1 root root 2225 Jul 21 2020 bzdiff + lrwxrwxrwx 1 root root 6 Jul 21 2020 bzegrep -> bzgrep + -rwxr-xr-x 1 root root 4877 Sep 5 2019 bzexe + lrwxrwxrwx 1 root root 6 Jul 21 2020 bzfgrep -> bzgrep + -rwxr-xr-x 1 root root 3775 Jul 21 2020 bzgrep + -rwxr-xr-x 3 root root 34808 Jul 21 2020 bzip2 + -rwxr-xr-x 1 root root 14264 Jul 21 2020 bzip2recover + lrwxrwxrwx 1 root root 6 Jul 21 2020 bzless -> bzmore + -rwxr-xr-x 1 root root 1297 Jul 21 2020 bzmore + -rwxr-xr-x 1 root root 39832 Sep 23 2020 cat + -rwxr-xr-x 1 root root 64512 Sep 23 2020 chgrp + -rwxr-xr-x 1 root root 60368 Sep 23 2020 chmod + -rwxr-xr-x 1 root root 64528 Sep 23 2020 chown + -rwxr-xr-x 1 root root 138896 Sep 23 2020 cp + -rwxr-xr-x 1 root root 129544 Dec 11 2020 dash + -rwxr-xr-x 1 root root 101384 Sep 23 2020 date + -rwxr-xr-x 1 root root 80984 Sep 23 2020 dd + -rwxr-xr-x 1 root root 89824 Sep 23 2020 df + -rwxr-xr-x 1 root root 143088 Sep 23 2020 dir + -rwxr-xr-x 1 root root 76152 Feb 8 04:38 dmesg + lrwxrwxrwx 1 root root 8 Nov 8 2019 dnsdomainname -> hostname + lrwxrwxrwx 1 root root 8 Nov 8 2019 domainname -> hostname + -rwxr-xr-x 1 root root 35632 Sep 23 2020 echo + -rwxr-xr-x 1 root root 28 Nov 10 2020 egrep + -rwxr-xr-x 1 root root 31512 Sep 23 2020 false + -rwxr-xr-x 1 root root 28 Nov 10 2020 fgrep + -rwxr-xr-x 1 root root 64856 Feb 8 04:38 findmnt + -rwsr-xr-x 1 root root 34824 Feb 27 06:12 fusermount + -rwxr-xr-x 1 root root 178400 Nov 10 2020 grep + -rwxr-xr-x 2 root root 2346 Mar 3 13:30 gunzip + -rwxr-xr-x 1 root root 6376 Mar 3 13:30 gzexe + -rwxr-xr-x 1 root root 93744 Mar 3 13:30 gzip + -rwxr-xr-x 1 root root 18440 Nov 8 2019 hostname + -rwxr-xr-x 1 root root 68720 Sep 23 2020 ln + -rwxr-xr-x 1 root root 52720 Feb 8 2020 login + -rwxr-xr-x 1 root root 143088 Sep 23 2020 ls + -rwxr-xr-x 1 root root 161960 Feb 8 04:38 lsblk + -rwxr-xr-x 1 root root 85200 Sep 23 2020 mkdir + -rwxr-xr-x 1 root root 68744 Sep 23 2020 mknod + -rwxr-xr-x 1 root root 43976 Sep 23 2020 mktemp + -rwxr-xr-x 1 root root 51368 Feb 8 04:38 more + -rwsr-xr-x 1 root root 51360 Feb 8 04:38 mount + -rwxr-xr-x 1 root root 14496 Feb 8 04:38 mountpoint + -rwxr-xr-x 1 root root 134808 Sep 23 2020 mv + lrwxrwxrwx 1 root root 8 Nov 8 2019 nisdomainname -> hostname + lrwxrwxrwx 1 root root 14 Apr 19 05:38 pidof -> /sbin/killall5 + -rwxr-xr-x 1 root root 35720 Sep 23 2020 pwd + lrwxrwxrwx 1 root root 4 Jun 22 16:26 rbash -> bash + -rwxr-xr-x 1 root root 43872 Sep 23 2020 readlink + -rwxr-xr-x 1 root root 68592 Sep 23 2020 rm + -rwxr-xr-x 1 root root 43880 Sep 23 2020 rmdir + -rwxr-xr-x 1 root root 19208 Sep 28 2020 run-parts + -rwxr-xr-x 1 root root 114016 Dec 23 2018 sed + lrwxrwxrwx 1 root root 4 Jul 18 09:24 sh -> bash + lrwxrwxrwx 1 root root 4 Jul 16 23:24 sh.distrib -> dash + -rwxr-xr-x 1 root root 35656 Sep 23 2020 sleep + -rwxr-xr-x 1 root root 72640 Sep 23 2020 stty + -rwsr-xr-x 1 root root 67776 Feb 8 04:38 su + -rwxr-xr-x 1 root root 35672 Sep 23 2020 sync + -rwxr-xr-x 1 root root 535768 Feb 17 23:55 tar + -rwxr-xr-x 1 root root 10568 Sep 28 2020 tempfile + -rwxr-xr-x 1 root root 89120 Sep 23 2020 touch + -rwxr-xr-x 1 root root 31512 Sep 23 2020 true + -rwxr-xr-x 1 root root 14264 Feb 27 06:12 ulockmgr_server + -rwsr-xr-x 1 root root 30880 Feb 8 04:38 umount + -rwxr-xr-x 1 root root 35640 Sep 23 2020 uname + -rwxr-xr-x 2 root root 2346 Mar 3 13:30 uncompress + -rwxr-xr-x 1 root root 143088 Sep 23 2020 vdir + -rwxr-xr-x 1 root root 59584 Feb 8 04:38 wdctl + lrwxrwxrwx 1 root root 8 Nov 8 2019 ypdomainname -> hostname + -rwxr-xr-x 1 root root 1984 Mar 3 13:30 zcat + -rwxr-xr-x 1 root root 1678 Mar 3 13:30 zcmp + -rwxr-xr-x 1 root root 5880 Mar 3 13:30 zdiff + -rwxr-xr-x 1 root root 29 Mar 3 13:30 zegrep + -rwxr-xr-x 1 root root 29 Mar 3 13:30 zfgrep + -rwxr-xr-x 1 root root 2081 Mar 3 13:30 zforce + -rwxr-xr-x 1 root root 7585 Mar 3 13:30 zgrep + -rwxr-xr-x 1 root root 2206 Mar 3 13:30 zless + -rwxr-xr-x 1 root root 1842 Mar 3 13:30 zmore + -rwxr-xr-x 1 root root 4553 Mar 3 13:30 znew +I: user script /srv/workspace/pbuilder/3330/tmp/hooks/D02_print_environment finished -> Attempting to satisfy build-dependencies -> Creating pbuilder-satisfydepends-dummy package Package: pbuilder-satisfydepends-dummy @@ -835,7 +869,7 @@ Get: 550 http://deb.debian.org/debian bullseye/main arm64 python3-tifffile all 20210201-1 [134 kB] Get: 551 http://deb.debian.org/debian bullseye/main arm64 python3-skimage all 0.18.1-2 [20.6 MB] Get: 552 http://deb.debian.org/debian bullseye/main arm64 python3-yaml arm64 5.3.1-5 [118 kB] -Fetched 530 MB in 10s (55.0 MB/s) +Fetched 530 MB in 9s (61.3 MB/s) debconf: delaying package configuration, since apt-utils is not installed Selecting previously unselected package libapparmor1:arm64. (Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 19646 files and directories currently installed.) @@ -3103,9 +3137,42 @@ Building dependency tree... Reading state information... fakeroot is already the newest version (1.25.3-1.1). -0 upgraded, 0 newly installed, 0 to remove and 0 not upgraded. +The following additional packages will be installed: + libfile-find-rule-perl libnumber-compare-perl libtext-glob-perl +The following NEW packages will be installed: + libfile-find-rule-perl libnumber-compare-perl libtext-glob-perl usrmerge +0 upgraded, 4 newly installed, 0 to remove and 0 not upgraded. +Need to get 59.5 kB of archives. +After this operation, 157 kB of additional disk space will be used. +Get:1 http://deb.debian.org/debian bullseye/main arm64 libnumber-compare-perl all 0.03-1.1 [6956 B] +Get:2 http://deb.debian.org/debian bullseye/main arm64 libtext-glob-perl all 0.11-1 [8888 B] +Get:3 http://deb.debian.org/debian bullseye/main arm64 libfile-find-rule-perl all 0.34-1 [30.6 kB] +Get:4 http://deb.debian.org/debian bullseye/main arm64 usrmerge all 25 [13.0 kB] +debconf: delaying package configuration, since apt-utils is not installed +Fetched 59.5 kB in 1s (56.3 kB/s) +Selecting previously unselected package libnumber-compare-perl. +(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 100932 files and directories currently installed.) +Preparing to unpack .../libnumber-compare-perl_0.03-1.1_all.deb ... +Unpacking libnumber-compare-perl (0.03-1.1) ... +Selecting previously unselected package libtext-glob-perl. +Preparing to unpack .../libtext-glob-perl_0.11-1_all.deb ... +Unpacking libtext-glob-perl (0.11-1) ... +Selecting previously unselected package libfile-find-rule-perl. +Preparing to unpack .../libfile-find-rule-perl_0.34-1_all.deb ... +Unpacking libfile-find-rule-perl (0.34-1) ... +Selecting previously unselected package usrmerge. +Preparing to unpack .../archives/usrmerge_25_all.deb ... +Unpacking usrmerge (25) ... +Setting up libtext-glob-perl (0.11-1) ... +Setting up libnumber-compare-perl (0.03-1.1) ... +Setting up libfile-find-rule-perl (0.34-1) ... +Setting up usrmerge (25) ... +The system has been successfully converted. +Processing triggers for man-db (2.9.4-2) ... +Not building database; man-db/auto-update is not 'true'. I: Building the package -I: Running cd /build/caffe-1.0.0+git20180821.99bd997/ && env PATH="/usr/sbin:/usr/bin:/sbin:/bin:/usr/games" HOME="/nonexistent/first-build" dpkg-buildpackage -us -uc -b && env PATH="/usr/sbin:/usr/bin:/sbin:/bin:/usr/games" HOME="/nonexistent/first-build" dpkg-genchanges -S > ../caffe_1.0.0+git20180821.99bd997-8_source.changes +hostname: Temporary failure in name resolution +I: Running cd /build/caffe-1.0.0+git20180821.99bd997/ && env PATH="/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/i/capture/the/path" HOME="/nonexistent/second-build" dpkg-buildpackage -us -uc -b && env PATH="/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/i/capture/the/path" HOME="/nonexistent/second-build" dpkg-genchanges -S > ../caffe_1.0.0+git20180821.99bd997-8_source.changes dpkg-buildpackage: info: source package caffe dpkg-buildpackage: info: source version 1.0.0+git20180821.99bd997-8 dpkg-buildpackage: info: source distribution unstable @@ -3282,22 +3349,22 @@ make[5]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' make -f src/caffe/CMakeFiles/caffe.dir/build.make src/caffe/CMakeFiles/caffe.dir/build make[5]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' -[ 1%] Building CXX object src/caffe/CMakeFiles/caffe.dir/common.cpp.o [ 1%] Building CXX object src/caffe/CMakeFiles/caffe.dir/blob.cpp.o +[ 1%] Building CXX object src/caffe/CMakeFiles/caffe.dir/common.cpp.o +[ 1%] Building CXX object src/caffe/CMakeFiles/caffe.dir/internal_thread.cpp.o [ 3%] Building CXX object src/caffe/CMakeFiles/caffe.dir/data_transformer.cpp.o cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -DWITH_PYTHON_LAYER -Dcaffe_EXPORTS -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/usr/include/python3.9 -I/usr/lib/python3/dist-packages/numpy/core/include -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -fPIC -o CMakeFiles/caffe.dir/blob.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/blob.cpp cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -DWITH_PYTHON_LAYER -Dcaffe_EXPORTS -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/usr/include/python3.9 -I/usr/lib/python3/dist-packages/numpy/core/include -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -fPIC -o CMakeFiles/caffe.dir/common.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/common.cpp -[ 5%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/absval_layer.cpp.o -[ 7%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layer.cpp.o cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -DWITH_PYTHON_LAYER -Dcaffe_EXPORTS -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/usr/include/python3.9 -I/usr/lib/python3/dist-packages/numpy/core/include -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -fPIC -o CMakeFiles/caffe.dir/data_transformer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/data_transformer.cpp -[ 7%] Building CXX object src/caffe/CMakeFiles/caffe.dir/internal_thread.cpp.o +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -DWITH_PYTHON_LAYER -Dcaffe_EXPORTS -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/usr/include/python3.9 -I/usr/lib/python3/dist-packages/numpy/core/include -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -fPIC -o CMakeFiles/caffe.dir/internal_thread.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/internal_thread.cpp +[ 5%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layer.cpp.o cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -DWITH_PYTHON_LAYER -Dcaffe_EXPORTS -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/usr/include/python3.9 -I/usr/lib/python3/dist-packages/numpy/core/include -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -fPIC -o CMakeFiles/caffe.dir/layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/layer.cpp +[ 5%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/accuracy_layer.cpp.o +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -DWITH_PYTHON_LAYER -Dcaffe_EXPORTS -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/usr/include/python3.9 -I/usr/lib/python3/dist-packages/numpy/core/include -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -fPIC -o CMakeFiles/caffe.dir/layers/accuracy_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/layers/accuracy_layer.cpp [ 7%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layer_factory.cpp.o -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -DWITH_PYTHON_LAYER -Dcaffe_EXPORTS -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/usr/include/python3.9 -I/usr/lib/python3/dist-packages/numpy/core/include -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -fPIC -o CMakeFiles/caffe.dir/layers/absval_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/layers/absval_layer.cpp -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -DWITH_PYTHON_LAYER -Dcaffe_EXPORTS -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/usr/include/python3.9 -I/usr/lib/python3/dist-packages/numpy/core/include -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -fPIC -o CMakeFiles/caffe.dir/internal_thread.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/internal_thread.cpp +[ 7%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/absval_layer.cpp.o cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -DWITH_PYTHON_LAYER -Dcaffe_EXPORTS -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/usr/include/python3.9 -I/usr/lib/python3/dist-packages/numpy/core/include -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -fPIC -o CMakeFiles/caffe.dir/layer_factory.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/layer_factory.cpp -[ 7%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/accuracy_layer.cpp.o -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -DWITH_PYTHON_LAYER -Dcaffe_EXPORTS -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/usr/include/python3.9 -I/usr/lib/python3/dist-packages/numpy/core/include -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -fPIC -o CMakeFiles/caffe.dir/layers/accuracy_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/layers/accuracy_layer.cpp +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -DWITH_PYTHON_LAYER -Dcaffe_EXPORTS -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/usr/include/python3.9 -I/usr/lib/python3/dist-packages/numpy/core/include -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -fPIC -o CMakeFiles/caffe.dir/layers/absval_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/layers/absval_layer.cpp In file included from /usr/include/boost/smart_ptr/detail/sp_thread_sleep.hpp:22, from /usr/include/boost/smart_ptr/detail/yield_k.hpp:23, from /usr/include/boost/smart_ptr/detail/spinlock_gcc_atomic.hpp:14, @@ -3322,8 +3389,6 @@ cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -DWITH_PYTHON_LAYER -Dcaffe_EXPORTS -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/usr/include/python3.9 -I/usr/lib/python3/dist-packages/numpy/core/include -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -fPIC -o CMakeFiles/caffe.dir/layers/argmax_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/layers/argmax_layer.cpp [ 9%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/base_conv_layer.cpp.o cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -DWITH_PYTHON_LAYER -Dcaffe_EXPORTS -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/usr/include/python3.9 -I/usr/lib/python3/dist-packages/numpy/core/include -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -fPIC -o CMakeFiles/caffe.dir/layers/base_conv_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/layers/base_conv_layer.cpp -[ 11%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/base_data_layer.cpp.o -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -DWITH_PYTHON_LAYER -Dcaffe_EXPORTS -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/usr/include/python3.9 -I/usr/lib/python3/dist-packages/numpy/core/include -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -fPIC -o CMakeFiles/caffe.dir/layers/base_data_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/layers/base_data_layer.cpp /build/caffe-1.0.0+git20180821.99bd997/src/caffe/layer_factory.cpp: In instantiation of 'boost::shared_ptr > caffe::GetPythonLayer(const caffe::LayerParameter&) [with Dtype = float]': /build/caffe-1.0.0+git20180821.99bd997/src/caffe/layer_factory.cpp:304:1: required from here /build/caffe-1.0.0+git20180821.99bd997/src/caffe/layer_factory.cpp:298:16: warning: catching polymorphic type 'struct boost::python::error_already_set' by value [-Wcatch-value=] @@ -3332,6 +3397,8 @@ /build/caffe-1.0.0+git20180821.99bd997/src/caffe/layer_factory.cpp: In instantiation of 'boost::shared_ptr > caffe::GetPythonLayer(const caffe::LayerParameter&) [with Dtype = double]': /build/caffe-1.0.0+git20180821.99bd997/src/caffe/layer_factory.cpp:304:1: required from here /build/caffe-1.0.0+git20180821.99bd997/src/caffe/layer_factory.cpp:298:16: warning: catching polymorphic type 'struct boost::python::error_already_set' by value [-Wcatch-value=] +[ 11%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/base_data_layer.cpp.o +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -DWITH_PYTHON_LAYER -Dcaffe_EXPORTS -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/usr/include/python3.9 -I/usr/lib/python3/dist-packages/numpy/core/include -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -fPIC -o CMakeFiles/caffe.dir/layers/base_data_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/layers/base_data_layer.cpp [ 11%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/batch_norm_layer.cpp.o cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -DWITH_PYTHON_LAYER -Dcaffe_EXPORTS -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/usr/include/python3.9 -I/usr/lib/python3/dist-packages/numpy/core/include -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -fPIC -o CMakeFiles/caffe.dir/layers/batch_norm_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/layers/batch_norm_layer.cpp [ 13%] Building CXX object src/caffe/CMakeFiles/caffe.dir/layers/batch_reindex_layer.cpp.o @@ -3667,21 +3734,21 @@ make[5]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' make -f src/caffe/test/CMakeFiles/test.testbin.dir/build.make src/caffe/test/CMakeFiles/test.testbin.dir/build make[5]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' -[ 62%] Building CXX object src/caffe/test/CMakeFiles/test.testbin.dir/test_batch_reindex_layer.cpp.o -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe/test && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DGTEST_USE_OWN_TR1_TUPLE -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/src -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/test.testbin.dir/test_batch_reindex_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/test/test_batch_reindex_layer.cpp -[ 63%] Building CXX object src/caffe/test/CMakeFiles/test.testbin.dir/test_argmax_layer.cpp.o [ 63%] Building CXX object src/caffe/test/CMakeFiles/test.testbin.dir/test_accuracy_layer.cpp.o -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe/test && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DGTEST_USE_OWN_TR1_TUPLE -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/src -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/test.testbin.dir/test_argmax_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/test/test_argmax_layer.cpp -[ 63%] Building CXX object src/caffe/test/CMakeFiles/test.testbin.dir/test_batch_norm_layer.cpp.o -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe/test && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DGTEST_USE_OWN_TR1_TUPLE -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/src -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/test.testbin.dir/test_accuracy_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/test/test_accuracy_layer.cpp -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe/test && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DGTEST_USE_OWN_TR1_TUPLE -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/src -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/test.testbin.dir/test_batch_norm_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/test/test_batch_norm_layer.cpp +[ 63%] Building CXX object src/caffe/test/CMakeFiles/test.testbin.dir/test_argmax_layer.cpp.o +[ 63%] Building CXX object src/caffe/test/CMakeFiles/test.testbin.dir/test_benchmark.cpp.o +[ 63%] Building CXX object src/caffe/test/CMakeFiles/test.testbin.dir/test_batch_reindex_layer.cpp.o [ 64%] Building CXX object src/caffe/test/CMakeFiles/test.testbin.dir/test_bias_layer.cpp.o -[ 64%] Building CXX object src/caffe/test/CMakeFiles/test.testbin.dir/test_benchmark.cpp.o +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe/test && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DGTEST_USE_OWN_TR1_TUPLE -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/src -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/test.testbin.dir/test_accuracy_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/test/test_accuracy_layer.cpp +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe/test && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DGTEST_USE_OWN_TR1_TUPLE -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/src -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/test.testbin.dir/test_argmax_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/test/test_argmax_layer.cpp +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe/test && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DGTEST_USE_OWN_TR1_TUPLE -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/src -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/test.testbin.dir/test_batch_reindex_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/test/test_batch_reindex_layer.cpp +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe/test && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DGTEST_USE_OWN_TR1_TUPLE -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/src -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/test.testbin.dir/test_benchmark.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/test/test_benchmark.cpp cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe/test && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DGTEST_USE_OWN_TR1_TUPLE -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/src -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/test.testbin.dir/test_bias_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/test/test_bias_layer.cpp [ 64%] Building CXX object src/caffe/test/CMakeFiles/test.testbin.dir/test_caffe_main.cpp.o -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe/test && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DGTEST_USE_OWN_TR1_TUPLE -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/src -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/test.testbin.dir/test_benchmark.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/test/test_benchmark.cpp -[ 64%] Building CXX object src/caffe/test/CMakeFiles/test.testbin.dir/test_blob.cpp.o cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe/test && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DGTEST_USE_OWN_TR1_TUPLE -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/src -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/test.testbin.dir/test_caffe_main.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/test/test_caffe_main.cpp +[ 64%] Building CXX object src/caffe/test/CMakeFiles/test.testbin.dir/test_batch_norm_layer.cpp.o +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe/test && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DGTEST_USE_OWN_TR1_TUPLE -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/src -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/test.testbin.dir/test_batch_norm_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/test/test_batch_norm_layer.cpp +[ 64%] Building CXX object src/caffe/test/CMakeFiles/test.testbin.dir/test_blob.cpp.o cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe/test && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DGTEST_USE_OWN_TR1_TUPLE -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/src -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/test.testbin.dir/test_blob.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/test/test_blob.cpp [ 65%] Building CXX object src/caffe/test/CMakeFiles/test.testbin.dir/test_common.cpp.o cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe/test && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DGTEST_USE_OWN_TR1_TUPLE -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/src -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/test.testbin.dir/test_common.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/test/test_common.cpp @@ -3790,8 +3857,8 @@ [ 95%] Building CXX object src/caffe/test/CMakeFiles/test.testbin.dir/test_spp_layer.cpp.o cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe/test && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DGTEST_USE_OWN_TR1_TUPLE -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/src -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/test.testbin.dir/test_spp_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/test/test_spp_layer.cpp [ 95%] Building CXX object src/caffe/test/CMakeFiles/test.testbin.dir/test_stochastic_pooling.cpp.o -[ 96%] Building CXX object src/caffe/test/CMakeFiles/test.testbin.dir/test_syncedmem.cpp.o cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe/test && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DGTEST_USE_OWN_TR1_TUPLE -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/src -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/test.testbin.dir/test_stochastic_pooling.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/test/test_stochastic_pooling.cpp +[ 96%] Building CXX object src/caffe/test/CMakeFiles/test.testbin.dir/test_syncedmem.cpp.o cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe/test && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DGTEST_USE_OWN_TR1_TUPLE -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/src -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/test.testbin.dir/test_syncedmem.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/test/test_syncedmem.cpp [ 96%] Building CXX object src/caffe/test/CMakeFiles/test.testbin.dir/test_tanh_layer.cpp.o cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe/test && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DGTEST_USE_OWN_TR1_TUPLE -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/src -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/test.testbin.dir/test_tanh_layer.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/src/caffe/test/test_tanh_layer.cpp @@ -3839,93 +3906,90 @@ make -f tools/CMakeFiles/compute_image_mean.dir/build.make tools/CMakeFiles/compute_image_mean.dir/depend make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build && /usr/bin/cmake -E cmake_depends "Unix Makefiles" /build/caffe-1.0.0+git20180821.99bd997 /build/caffe-1.0.0+git20180821.99bd997/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/upgrade_net_proto_text.dir/DependInfo.cmake --color= +make -f tools/CMakeFiles/upgrade_solver_proto_text.dir/build.make tools/CMakeFiles/upgrade_solver_proto_text.dir/depend make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build && /usr/bin/cmake -E cmake_depends "Unix Makefiles" /build/caffe-1.0.0+git20180821.99bd997 /build/caffe-1.0.0+git20180821.99bd997/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/upgrade_net_proto_binary.dir/DependInfo.cmake --color= -make -f tools/CMakeFiles/upgrade_solver_proto_text.dir/build.make tools/CMakeFiles/upgrade_solver_proto_text.dir/depend make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build && /usr/bin/cmake -E cmake_depends "Unix Makefiles" /build/caffe-1.0.0+git20180821.99bd997 /build/caffe-1.0.0+git20180821.99bd997/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/extract_features.dir/DependInfo.cmake --color= -make -f tools/CMakeFiles/caffe.bin.dir/build.make tools/CMakeFiles/caffe.bin.dir/depend make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build && /usr/bin/cmake -E cmake_depends "Unix Makefiles" /build/caffe-1.0.0+git20180821.99bd997 /build/caffe-1.0.0+git20180821.99bd997/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/convert_imageset.dir/DependInfo.cmake --color= make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build && /usr/bin/cmake -E cmake_depends "Unix Makefiles" /build/caffe-1.0.0+git20180821.99bd997 /build/caffe-1.0.0+git20180821.99bd997/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/upgrade_solver_proto_text.dir/DependInfo.cmake --color= +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build && /usr/bin/cmake -E cmake_depends "Unix Makefiles" /build/caffe-1.0.0+git20180821.99bd997 /build/caffe-1.0.0+git20180821.99bd997/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/compute_image_mean.dir/DependInfo.cmake --color= make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build && /usr/bin/cmake -E cmake_depends "Unix Makefiles" /build/caffe-1.0.0+git20180821.99bd997 /build/caffe-1.0.0+git20180821.99bd997/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/caffe.bin.dir/DependInfo.cmake --color= +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build && /usr/bin/cmake -E cmake_depends "Unix Makefiles" /build/caffe-1.0.0+git20180821.99bd997 /build/caffe-1.0.0+git20180821.99bd997/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/upgrade_solver_proto_text.dir/DependInfo.cmake --color= +make -f tools/CMakeFiles/caffe.bin.dir/build.make tools/CMakeFiles/caffe.bin.dir/depend make -f examples/CMakeFiles/convert_mnist_siamese_data.dir/build.make examples/CMakeFiles/convert_mnist_siamese_data.dir/depend make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build && /usr/bin/cmake -E cmake_depends "Unix Makefiles" /build/caffe-1.0.0+git20180821.99bd997 /build/caffe-1.0.0+git20180821.99bd997/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/compute_image_mean.dir/DependInfo.cmake --color= +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build && /usr/bin/cmake -E cmake_depends "Unix Makefiles" /build/caffe-1.0.0+git20180821.99bd997 /build/caffe-1.0.0+git20180821.99bd997/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/caffe.bin.dir/DependInfo.cmake --color= make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build && /usr/bin/cmake -E cmake_depends "Unix Makefiles" /build/caffe-1.0.0+git20180821.99bd997 /build/caffe-1.0.0+git20180821.99bd997/examples /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples/CMakeFiles/convert_mnist_siamese_data.dir/DependInfo.cmake --color= -Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/caffe.bin.dir/DependInfo.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/caffe.bin.dir/depend.internal". -Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/CMakeDirectoryInformation.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/caffe.bin.dir/depend.internal". -Scanning dependencies of target caffe.bin -Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/upgrade_net_proto_text.dir/DependInfo.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/upgrade_net_proto_text.dir/depend.internal". -Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/CMakeDirectoryInformation.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/upgrade_net_proto_text.dir/depend.internal". -make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' -make -f tools/CMakeFiles/caffe.bin.dir/build.make tools/CMakeFiles/caffe.bin.dir/build -make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/extract_features.dir/DependInfo.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/extract_features.dir/depend.internal". +Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/convert_imageset.dir/DependInfo.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/convert_imageset.dir/depend.internal". Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/upgrade_net_proto_binary.dir/DependInfo.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/upgrade_net_proto_binary.dir/depend.internal". -Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/CMakeDirectoryInformation.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/extract_features.dir/depend.internal". +Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/compute_image_mean.dir/DependInfo.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/compute_image_mean.dir/depend.internal". +Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/upgrade_net_proto_text.dir/DependInfo.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/upgrade_net_proto_text.dir/depend.internal". Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/CMakeDirectoryInformation.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/upgrade_net_proto_binary.dir/depend.internal". +Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/CMakeDirectoryInformation.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/convert_imageset.dir/depend.internal". +Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/CMakeDirectoryInformation.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/extract_features.dir/depend.internal". +Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/CMakeDirectoryInformation.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/compute_image_mean.dir/depend.internal". +Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/CMakeDirectoryInformation.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/upgrade_net_proto_text.dir/depend.internal". Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/upgrade_solver_proto_text.dir/DependInfo.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/upgrade_solver_proto_text.dir/depend.internal". Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/CMakeDirectoryInformation.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/upgrade_solver_proto_text.dir/depend.internal". -[ 81%] Building CXX object tools/CMakeFiles/caffe.bin.dir/caffe.cpp.o +Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/caffe.bin.dir/DependInfo.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/caffe.bin.dir/depend.internal". +Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/CMakeDirectoryInformation.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/caffe.bin.dir/depend.internal". +Scanning dependencies of target convert_imageset Scanning dependencies of target extract_features -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/caffe.bin.dir/caffe.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/tools/caffe.cpp -Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/convert_imageset.dir/DependInfo.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/convert_imageset.dir/depend.internal". -Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/CMakeDirectoryInformation.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/convert_imageset.dir/depend.internal". -make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' -Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/compute_image_mean.dir/DependInfo.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/compute_image_mean.dir/depend.internal". -Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/CMakeDirectoryInformation.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools/CMakeFiles/compute_image_mean.dir/depend.internal". -Scanning dependencies of target upgrade_net_proto_binary -make -f tools/CMakeFiles/extract_features.dir/build.make tools/CMakeFiles/extract_features.dir/build +Scanning dependencies of target compute_image_mean make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' -make -f tools/CMakeFiles/upgrade_net_proto_binary.dir/build.make tools/CMakeFiles/upgrade_net_proto_binary.dir/build +make -f tools/CMakeFiles/convert_imageset.dir/build.make tools/CMakeFiles/convert_imageset.dir/build Scanning dependencies of target upgrade_solver_proto_text -Scanning dependencies of target convert_imageset +make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' -make -f tools/CMakeFiles/upgrade_solver_proto_text.dir/build.make tools/CMakeFiles/upgrade_solver_proto_text.dir/build +make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' +make -f tools/CMakeFiles/extract_features.dir/build.make tools/CMakeFiles/extract_features.dir/build +make -f tools/CMakeFiles/compute_image_mean.dir/build.make tools/CMakeFiles/compute_image_mean.dir/build +Scanning dependencies of target upgrade_net_proto_text +Scanning dependencies of target upgrade_net_proto_binary make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' -make -f tools/CMakeFiles/convert_imageset.dir/build.make tools/CMakeFiles/convert_imageset.dir/build +make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' +make -f tools/CMakeFiles/upgrade_net_proto_text.dir/build.make tools/CMakeFiles/upgrade_net_proto_text.dir/build +make -f tools/CMakeFiles/upgrade_net_proto_binary.dir/build.make tools/CMakeFiles/upgrade_net_proto_binary.dir/build +make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples/CMakeFiles/convert_mnist_siamese_data.dir/DependInfo.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples/CMakeFiles/convert_mnist_siamese_data.dir/depend.internal". Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples/CMakeFiles/CMakeDirectoryInformation.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples/CMakeFiles/convert_mnist_siamese_data.dir/depend.internal". -make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' -[ 82%] Building CXX object tools/CMakeFiles/extract_features.dir/extract_features.cpp.o -Scanning dependencies of target upgrade_net_proto_text -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/extract_features.dir/extract_features.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/tools/extract_features.cpp +Scanning dependencies of target caffe.bin +make -f tools/CMakeFiles/upgrade_solver_proto_text.dir/build.make tools/CMakeFiles/upgrade_solver_proto_text.dir/build make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' -[ 82%] Building CXX object tools/CMakeFiles/upgrade_net_proto_binary.dir/upgrade_net_proto_binary.cpp.o -make -f tools/CMakeFiles/upgrade_net_proto_text.dir/build.make tools/CMakeFiles/upgrade_net_proto_text.dir/build +make -f tools/CMakeFiles/caffe.bin.dir/build.make tools/CMakeFiles/caffe.bin.dir/build +make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' +make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' +[ 82%] Building CXX object tools/CMakeFiles/convert_imageset.dir/convert_imageset.cpp.o +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/convert_imageset.dir/convert_imageset.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/tools/convert_imageset.cpp +[ 82%] Building CXX object tools/CMakeFiles/upgrade_net_proto_text.dir/upgrade_net_proto_text.cpp.o Scanning dependencies of target convert_mnist_siamese_data -Scanning dependencies of target compute_image_mean -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/upgrade_net_proto_binary.dir/upgrade_net_proto_binary.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/tools/upgrade_net_proto_binary.cpp +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/upgrade_net_proto_text.dir/upgrade_net_proto_text.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/tools/upgrade_net_proto_text.cpp make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' make -f examples/CMakeFiles/convert_mnist_siamese_data.dir/build.make examples/CMakeFiles/convert_mnist_siamese_data.dir/build -make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' -make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' -make -f tools/CMakeFiles/compute_image_mean.dir/build.make tools/CMakeFiles/compute_image_mean.dir/build make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' +[ 84%] Building CXX object tools/CMakeFiles/extract_features.dir/extract_features.cpp.o [ 84%] Building CXX object tools/CMakeFiles/upgrade_solver_proto_text.dir/upgrade_solver_proto_text.cpp.o -[ 84%] Building CXX object tools/CMakeFiles/convert_imageset.dir/convert_imageset.cpp.o -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/convert_imageset.dir/convert_imageset.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/tools/convert_imageset.cpp +[ 84%] Building CXX object tools/CMakeFiles/upgrade_net_proto_binary.dir/upgrade_net_proto_binary.cpp.o +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/extract_features.dir/extract_features.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/tools/extract_features.cpp cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/upgrade_solver_proto_text.dir/upgrade_solver_proto_text.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/tools/upgrade_solver_proto_text.cpp -[ 84%] Building CXX object tools/CMakeFiles/upgrade_net_proto_text.dir/upgrade_net_proto_text.cpp.o -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/upgrade_net_proto_text.dir/upgrade_net_proto_text.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/tools/upgrade_net_proto_text.cpp +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/upgrade_net_proto_binary.dir/upgrade_net_proto_binary.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/tools/upgrade_net_proto_binary.cpp [ 85%] Building CXX object tools/CMakeFiles/compute_image_mean.dir/compute_image_mean.cpp.o +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/compute_image_mean.dir/compute_image_mean.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/tools/compute_image_mean.cpp [ 87%] Building CXX object examples/CMakeFiles/convert_mnist_siamese_data.dir/siamese/convert_mnist_siamese_data.cpp.o cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/convert_mnist_siamese_data.dir/siamese/convert_mnist_siamese_data.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/examples/siamese/convert_mnist_siamese_data.cpp -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/compute_image_mean.dir/compute_image_mean.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/tools/compute_image_mean.cpp +[ 87%] Building CXX object tools/CMakeFiles/caffe.bin.dir/caffe.cpp.o +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/caffe.bin.dir/caffe.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/tools/caffe.cpp [ 87%] Linking CXX executable siamese/convert_mnist_siamese_data cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples && /usr/bin/cmake -E cmake_link_script CMakeFiles/convert_mnist_siamese_data.dir/link.txt --verbose=1 /usr/bin/c++ -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -Wl,-z,relro -Wl,-z,now -Wl,--as-needed -rdynamic CMakeFiles/convert_mnist_siamese_data.dir/siamese/convert_mnist_siamese_data.cpp.o -o siamese/convert_mnist_siamese_data ../lib/libcaffe.so.1.0.0 ../lib/libcaffeproto.a /usr/lib/aarch64-linux-gnu/libboost_system.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_thread.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_atomic.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_filesystem.so.1.74.0 -lglog -lgflags -lprotobuf -lpthread /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5.so -lpthread -lsz -lz -ldl -lm /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5.so -lpthread -lsz -lz -ldl -lm /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl.so -llmdb -lleveldb /usr/lib/aarch64-linux-gnu/libopencv_highgui.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_videoio.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_imgcodecs.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_imgproc.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_core.so.4.5.1 -lblas /usr/lib/aarch64-linux-gnu/libboost_python39.so.1.74.0 -[ 87%] Linking CXX executable compute_image_mean -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/cmake -E cmake_link_script CMakeFiles/compute_image_mean.dir/link.txt --verbose=1 -/usr/bin/c++ -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -Wl,-z,relro -Wl,-z,now -Wl,--as-needed -rdynamic CMakeFiles/compute_image_mean.dir/compute_image_mean.cpp.o -o compute_image_mean ../lib/libcaffe.so.1.0.0 ../lib/libcaffeproto.a /usr/lib/aarch64-linux-gnu/libboost_system.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_thread.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_atomic.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_filesystem.so.1.74.0 -lglog -lgflags -lprotobuf -lpthread /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5.so -lpthread -lsz -lz -ldl -lm /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5.so -lpthread -lsz -lz -ldl -lm /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl.so -llmdb -lleveldb /usr/lib/aarch64-linux-gnu/libopencv_highgui.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_videoio.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_imgcodecs.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_imgproc.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_core.so.4.5.1 -lblas /usr/lib/aarch64-linux-gnu/libboost_python39.so.1.74.0 cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples && ln -sf /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples/siamese/convert_mnist_siamese_data /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples/siamese/convert_mnist_siamese_data.bin make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' [ 87%] Built target convert_mnist_siamese_data @@ -3940,6 +4004,9 @@ make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' [ 89%] Building CXX object examples/CMakeFiles/convert_mnist_data.dir/mnist/convert_mnist_data.cpp.o cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/convert_mnist_data.dir/mnist/convert_mnist_data.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/examples/mnist/convert_mnist_data.cpp +[ 89%] Linking CXX executable compute_image_mean +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/cmake -E cmake_link_script CMakeFiles/compute_image_mean.dir/link.txt --verbose=1 +/usr/bin/c++ -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -Wl,-z,relro -Wl,-z,now -Wl,--as-needed -rdynamic CMakeFiles/compute_image_mean.dir/compute_image_mean.cpp.o -o compute_image_mean ../lib/libcaffe.so.1.0.0 ../lib/libcaffeproto.a /usr/lib/aarch64-linux-gnu/libboost_system.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_thread.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_atomic.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_filesystem.so.1.74.0 -lglog -lgflags -lprotobuf -lpthread /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5.so -lpthread -lsz -lz -ldl -lm /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5.so -lpthread -lsz -lz -ldl -lm /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl.so -llmdb -lleveldb /usr/lib/aarch64-linux-gnu/libopencv_highgui.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_videoio.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_imgcodecs.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_imgproc.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_core.so.4.5.1 -lblas /usr/lib/aarch64-linux-gnu/libboost_python39.so.1.74.0 make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' [ 89%] Built target compute_image_mean make -f examples/CMakeFiles/classification.dir/build.make examples/CMakeFiles/classification.dir/depend @@ -3959,25 +4026,25 @@ [ 92%] Linking CXX executable upgrade_net_proto_binary cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/cmake -E cmake_link_script CMakeFiles/upgrade_net_proto_binary.dir/link.txt --verbose=1 /usr/bin/c++ -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -Wl,-z,relro -Wl,-z,now -Wl,--as-needed -rdynamic CMakeFiles/upgrade_net_proto_binary.dir/upgrade_net_proto_binary.cpp.o -o upgrade_net_proto_binary ../lib/libcaffe.so.1.0.0 ../lib/libcaffeproto.a /usr/lib/aarch64-linux-gnu/libboost_system.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_thread.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_atomic.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_filesystem.so.1.74.0 -lglog -lgflags -lprotobuf -lpthread /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5.so -lpthread -lsz -lz -ldl -lm /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5.so -lpthread -lsz -lz -ldl -lm /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl.so -llmdb -lleveldb /usr/lib/aarch64-linux-gnu/libopencv_highgui.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_videoio.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_imgcodecs.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_imgproc.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_core.so.4.5.1 -lblas /usr/lib/aarch64-linux-gnu/libboost_python39.so.1.74.0 -[ 93%] Linking CXX executable upgrade_net_proto_text -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/cmake -E cmake_link_script CMakeFiles/upgrade_net_proto_text.dir/link.txt --verbose=1 -/usr/bin/c++ -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -Wl,-z,relro -Wl,-z,now -Wl,--as-needed -rdynamic CMakeFiles/upgrade_net_proto_text.dir/upgrade_net_proto_text.cpp.o -o upgrade_net_proto_text ../lib/libcaffe.so.1.0.0 ../lib/libcaffeproto.a /usr/lib/aarch64-linux-gnu/libboost_system.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_thread.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_atomic.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_filesystem.so.1.74.0 -lglog -lgflags -lprotobuf -lpthread /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5.so -lpthread -lsz -lz -ldl -lm /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5.so -lpthread -lsz -lz -ldl -lm /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl.so -llmdb -lleveldb /usr/lib/aarch64-linux-gnu/libopencv_highgui.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_videoio.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_imgcodecs.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_imgproc.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_core.so.4.5.1 -lblas /usr/lib/aarch64-linux-gnu/libboost_python39.so.1.74.0 +[ 93%] Linking CXX executable upgrade_solver_proto_text +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/cmake -E cmake_link_script CMakeFiles/upgrade_solver_proto_text.dir/link.txt --verbose=1 +/usr/bin/c++ -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -Wl,-z,relro -Wl,-z,now -Wl,--as-needed -rdynamic CMakeFiles/upgrade_solver_proto_text.dir/upgrade_solver_proto_text.cpp.o -o upgrade_solver_proto_text ../lib/libcaffe.so.1.0.0 ../lib/libcaffeproto.a /usr/lib/aarch64-linux-gnu/libboost_system.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_thread.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_atomic.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_filesystem.so.1.74.0 -lglog -lgflags -lprotobuf -lpthread /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5.so -lpthread -lsz -lz -ldl -lm /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5.so -lpthread -lsz -lz -ldl -lm /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl.so -llmdb -lleveldb /usr/lib/aarch64-linux-gnu/libopencv_highgui.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_videoio.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_imgcodecs.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_imgproc.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_core.so.4.5.1 -lblas /usr/lib/aarch64-linux-gnu/libboost_python39.so.1.74.0 make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' [ 93%] Built target convert_imageset make -f examples/CMakeFiles/convert_cifar_data.dir/build.make examples/CMakeFiles/convert_cifar_data.dir/depend make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build && /usr/bin/cmake -E cmake_depends "Unix Makefiles" /build/caffe-1.0.0+git20180821.99bd997 /build/caffe-1.0.0+git20180821.99bd997/examples /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples/CMakeFiles/convert_cifar_data.dir/DependInfo.cmake --color= +[ 95%] Linking CXX executable upgrade_net_proto_text Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples/CMakeFiles/convert_cifar_data.dir/DependInfo.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples/CMakeFiles/convert_cifar_data.dir/depend.internal". Dependee "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples/CMakeFiles/CMakeDirectoryInformation.cmake" is newer than depender "/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples/CMakeFiles/convert_cifar_data.dir/depend.internal". +cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/cmake -E cmake_link_script CMakeFiles/upgrade_net_proto_text.dir/link.txt --verbose=1 Scanning dependencies of target convert_cifar_data make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' make -f examples/CMakeFiles/convert_cifar_data.dir/build.make examples/CMakeFiles/convert_cifar_data.dir/build +/usr/bin/c++ -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -Wl,-z,relro -Wl,-z,now -Wl,--as-needed -rdynamic CMakeFiles/upgrade_net_proto_text.dir/upgrade_net_proto_text.cpp.o -o upgrade_net_proto_text ../lib/libcaffe.so.1.0.0 ../lib/libcaffeproto.a /usr/lib/aarch64-linux-gnu/libboost_system.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_thread.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_atomic.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_filesystem.so.1.74.0 -lglog -lgflags -lprotobuf -lpthread /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5.so -lpthread -lsz -lz -ldl -lm /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5.so -lpthread -lsz -lz -ldl -lm /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl.so -llmdb -lleveldb /usr/lib/aarch64-linux-gnu/libopencv_highgui.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_videoio.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_imgcodecs.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_imgproc.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_core.so.4.5.1 -lblas /usr/lib/aarch64-linux-gnu/libboost_python39.so.1.74.0 make[4]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' -[ 95%] Building CXX object examples/CMakeFiles/convert_cifar_data.dir/cifar10/convert_cifar_data.cpp.o +[ 96%] Building CXX object examples/CMakeFiles/convert_cifar_data.dir/cifar10/convert_cifar_data.cpp.o cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples && /usr/bin/c++ -DBOOST_ALL_NO_LIB -DBOOST_ATOMIC_DYN_LINK -DBOOST_FILESYSTEM_DYN_LINK -DBOOST_PYTHON_DYN_LINK -DBOOST_SYSTEM_DYN_LINK -DBOOST_THREAD_DYN_LINK -DCAFFE_VERSION=1.0.0 -DCPU_ONLY -DUSE_HDF5 -DUSE_LEVELDB -DUSE_LMDB -DUSE_OPENCV -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/include -I/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build -I/build/caffe-1.0.0+git20180821.99bd997/include -isystem /usr/include/hdf5/serial -isystem /usr/include/opencv4 -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -o CMakeFiles/convert_cifar_data.dir/cifar10/convert_cifar_data.cpp.o -c /build/caffe-1.0.0+git20180821.99bd997/examples/cifar10/convert_cifar_data.cpp -[ 96%] Linking CXX executable upgrade_solver_proto_text -cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/tools && /usr/bin/cmake -E cmake_link_script CMakeFiles/upgrade_solver_proto_text.dir/link.txt --verbose=1 -/usr/bin/c++ -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -Wl,-z,relro -Wl,-z,now -Wl,--as-needed -rdynamic CMakeFiles/upgrade_solver_proto_text.dir/upgrade_solver_proto_text.cpp.o -o upgrade_solver_proto_text ../lib/libcaffe.so.1.0.0 ../lib/libcaffeproto.a /usr/lib/aarch64-linux-gnu/libboost_system.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_thread.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_atomic.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_filesystem.so.1.74.0 -lglog -lgflags -lprotobuf -lpthread /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5.so -lpthread -lsz -lz -ldl -lm /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5.so -lpthread -lsz -lz -ldl -lm /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl.so -llmdb -lleveldb /usr/lib/aarch64-linux-gnu/libopencv_highgui.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_videoio.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_imgcodecs.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_imgproc.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_core.so.4.5.1 -lblas /usr/lib/aarch64-linux-gnu/libboost_python39.so.1.74.0 make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' [ 96%] Built target upgrade_net_proto_binary make -f python/CMakeFiles/pycaffe.dir/build.make python/CMakeFiles/pycaffe.dir/depend @@ -3990,9 +4057,9 @@ make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' [ 98%] Built target pycaffe make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' -[ 98%] Built target upgrade_net_proto_text -make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' [ 98%] Built target upgrade_solver_proto_text +make[4]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' +[ 98%] Built target upgrade_net_proto_text [ 98%] Linking CXX executable mnist/convert_mnist_data cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/examples && /usr/bin/cmake -E cmake_link_script CMakeFiles/convert_mnist_data.dir/link.txt --verbose=1 /usr/bin/c++ -g -O2 -fdebug-prefix-map=/build/caffe-1.0.0+git20180821.99bd997=. -fstack-protector-strong -Wformat -Werror=format-security -Wall -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -Wall -Wno-sign-compare -Wno-uninitialized -O3 -DNDEBUG -Wl,-z,relro -Wl,-z,now -Wl,--as-needed -rdynamic CMakeFiles/convert_mnist_data.dir/mnist/convert_mnist_data.cpp.o -o mnist/convert_mnist_data ../lib/libcaffe.so.1.0.0 ../lib/libcaffeproto.a /usr/lib/aarch64-linux-gnu/libboost_system.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_thread.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_atomic.so.1.74.0 /usr/lib/aarch64-linux-gnu/libboost_filesystem.so.1.74.0 -lglog -lgflags -lprotobuf -lpthread /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5.so -lpthread -lsz -lz -ldl -lm /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5.so -lpthread -lsz -lz -ldl -lm /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl_cpp.so /usr/lib/aarch64-linux-gnu/hdf5/serial/libhdf5_hl.so -llmdb -lleveldb /usr/lib/aarch64-linux-gnu/libopencv_highgui.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_videoio.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_imgcodecs.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_imgproc.so.4.5.1 /usr/lib/aarch64-linux-gnu/libopencv_core.so.4.5.1 -lblas /usr/lib/aarch64-linux-gnu/libboost_python39.so.1.74.0 @@ -11935,20 +12002,20 @@ make[5]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build && /usr/bin/cmake -E cmake_depends "Unix Makefiles" /build/caffe-1.0.0+git20180821.99bd997 /build/caffe-1.0.0+git20180821.99bd997/src/gtest /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/gtest /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/gtest/CMakeFiles/gtest.dir/DependInfo.cmake --color= make[5]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' +make[5]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' make -f src/caffe/CMakeFiles/caffeproto.dir/build.make src/caffe/CMakeFiles/caffeproto.dir/build +make -f src/gtest/CMakeFiles/gtest.dir/build.make src/gtest/CMakeFiles/gtest.dir/build make[5]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' make[5]: Nothing to be done for 'src/caffe/CMakeFiles/caffeproto.dir/build'. make[5]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' -[ 1%] Built target caffeproto -make -f src/caffe/CMakeFiles/caffe.dir/build.make src/caffe/CMakeFiles/caffe.dir/depend -make[5]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' -make -f src/gtest/CMakeFiles/gtest.dir/build.make src/gtest/CMakeFiles/gtest.dir/build make[5]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' make[5]: Nothing to be done for 'src/gtest/CMakeFiles/gtest.dir/build'. make[5]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' +[ 1%] Built target caffeproto +make -f src/caffe/CMakeFiles/caffe.dir/build.make src/caffe/CMakeFiles/caffe.dir/depend +[ 2%] Built target gtest make[5]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' cd /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build && /usr/bin/cmake -E cmake_depends "Unix Makefiles" /build/caffe-1.0.0+git20180821.99bd997 /build/caffe-1.0.0+git20180821.99bd997/src/caffe /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/src/caffe/CMakeFiles/caffe.dir/DependInfo.cmake --color= -[ 2%] Built target gtest make[5]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' make -f src/caffe/CMakeFiles/caffe.dir/build.make src/caffe/CMakeFiles/caffe.dir/build make[5]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' @@ -11973,2798 +12040,2798 @@ make[5]: Entering directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' cd /build/caffe-1.0.0+git20180821.99bd997 && /build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build/test/test.testbin --gtest_shuffle --gtest_filter="-*GPU*" Note: Google Test filter = -*GPU* -Note: Randomizing tests' orders with a seed of 4140 . +Note: Randomizing tests' orders with a seed of 51699 . [==========] Running 1162 tests from 152 test cases. [----------] Global test environment set-up. +[----------] 4 tests from ConstantFillerTest/0, where TypeParam = float +[ RUN ] ConstantFillerTest/0.TestFill1D +[ OK ] ConstantFillerTest/0.TestFill1D (0 ms) +[ RUN ] ConstantFillerTest/0.TestFill +[ OK ] ConstantFillerTest/0.TestFill (0 ms) +[ RUN ] ConstantFillerTest/0.TestFill5D +[ OK ] ConstantFillerTest/0.TestFill5D (0 ms) +[ RUN ] ConstantFillerTest/0.TestFill2D +[ OK ] ConstantFillerTest/0.TestFill2D (0 ms) +[----------] 4 tests from ConstantFillerTest/0 (1 ms total) + +[----------] 12 tests from ArgMaxLayerTest/1, where TypeParam = double +[ RUN ] ArgMaxLayerTest/1.TestCPUAxis +[ OK ] ArgMaxLayerTest/1.TestCPUAxis (13 ms) +[ RUN ] ArgMaxLayerTest/1.TestSetup +[ OK ] ArgMaxLayerTest/1.TestSetup (2 ms) +[ RUN ] ArgMaxLayerTest/1.TestSetupAxisNegativeIndexing +[ OK ] ArgMaxLayerTest/1.TestSetupAxisNegativeIndexing (1 ms) +[ RUN ] ArgMaxLayerTest/1.TestCPU +[ OK ] ArgMaxLayerTest/1.TestCPU (3 ms) +[ RUN ] ArgMaxLayerTest/1.TestSetupMaxVal +[ OK ] ArgMaxLayerTest/1.TestSetupMaxVal (1 ms) +[ RUN ] ArgMaxLayerTest/1.TestCPUAxisMaxValTopK +[ OK ] ArgMaxLayerTest/1.TestCPUAxisMaxValTopK (51 ms) +[ RUN ] ArgMaxLayerTest/1.TestCPUTopK +[ OK ] ArgMaxLayerTest/1.TestCPUTopK (2 ms) +[ RUN ] ArgMaxLayerTest/1.TestSetupAxis +[ OK ] ArgMaxLayerTest/1.TestSetupAxis (2 ms) +[ RUN ] ArgMaxLayerTest/1.TestCPUMaxVal +[ OK ] ArgMaxLayerTest/1.TestCPUMaxVal (2 ms) +[ RUN ] ArgMaxLayerTest/1.TestSetupAxisMaxVal +[ OK ] ArgMaxLayerTest/1.TestSetupAxisMaxVal (1 ms) +[ RUN ] ArgMaxLayerTest/1.TestCPUMaxValTopK +[ OK ] ArgMaxLayerTest/1.TestCPUMaxValTopK (2 ms) +[ RUN ] ArgMaxLayerTest/1.TestCPUAxisTopK +[ OK ] ArgMaxLayerTest/1.TestCPUAxisTopK (54 ms) +[----------] 12 tests from ArgMaxLayerTest/1 (134 ms total) + +[----------] 4 tests from ConstantFillerTest/1, where TypeParam = double +[ RUN ] ConstantFillerTest/1.TestFill2D +[ OK ] ConstantFillerTest/1.TestFill2D (0 ms) +[ RUN ] ConstantFillerTest/1.TestFill5D +[ OK ] ConstantFillerTest/1.TestFill5D (0 ms) +[ RUN ] ConstantFillerTest/1.TestFill +[ OK ] ConstantFillerTest/1.TestFill (0 ms) +[ RUN ] ConstantFillerTest/1.TestFill1D +[ OK ] ConstantFillerTest/1.TestFill1D (0 ms) +[----------] 4 tests from ConstantFillerTest/1 (0 ms total) + +[----------] 8 tests from RMSPropSolverTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] RMSPropSolverTest/1.TestRMSPropLeastSquaresUpdateWithEverything +[ OK ] RMSPropSolverTest/1.TestRMSPropLeastSquaresUpdateWithEverything (125 ms) +[ RUN ] RMSPropSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare +[ OK ] RMSPropSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare (8 ms) +[ RUN ] RMSPropSolverTest/1.TestRMSPropLeastSquaresUpdateWithWeightDecay +[ OK ] RMSPropSolverTest/1.TestRMSPropLeastSquaresUpdateWithWeightDecay (23 ms) +[ RUN ] RMSPropSolverTest/1.TestSnapshotShare +[ OK ] RMSPropSolverTest/1.TestSnapshotShare (47 ms) +[ RUN ] RMSPropSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum +[ OK ] RMSPropSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum (6 ms) +[ RUN ] RMSPropSolverTest/1.TestSnapshot +[ OK ] RMSPropSolverTest/1.TestSnapshot (38 ms) +[ RUN ] RMSPropSolverTest/1.TestRMSPropLeastSquaresUpdateWithRmsDecay +[ OK ] RMSPropSolverTest/1.TestRMSPropLeastSquaresUpdateWithRmsDecay (115 ms) +[ RUN ] RMSPropSolverTest/1.TestRMSPropLeastSquaresUpdateWithEverythingShare +[ OK ] RMSPropSolverTest/1.TestRMSPropLeastSquaresUpdateWithEverythingShare (122 ms) +[----------] 8 tests from RMSPropSolverTest/1 (485 ms total) + +[----------] 2 tests from HingeLossLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] HingeLossLayerTest/1.TestGradientL1 +[ OK ] HingeLossLayerTest/1.TestGradientL1 (0 ms) +[ RUN ] HingeLossLayerTest/1.TestGradientL2 +[ OK ] HingeLossLayerTest/1.TestGradientL2 (1 ms) +[----------] 2 tests from HingeLossLayerTest/1 (1 ms total) + +[----------] 2 tests from BilinearFillerTest/0, where TypeParam = float +[ RUN ] BilinearFillerTest/0.TestFillEven +[ OK ] BilinearFillerTest/0.TestFillEven (12 ms) +[ RUN ] BilinearFillerTest/0.TestFillOdd +[ OK ] BilinearFillerTest/0.TestFillOdd (16 ms) +[----------] 2 tests from BilinearFillerTest/0 (28 ms total) + +[----------] 12 tests from DataTransformTest/0, where TypeParam = float +[ RUN ] DataTransformTest/0.TestCropSize +[ OK ] DataTransformTest/0.TestCropSize (0 ms) +[ RUN ] DataTransformTest/0.TestMirrorTest +[ OK ] DataTransformTest/0.TestMirrorTest (1 ms) +[ RUN ] DataTransformTest/0.TestEmptyTransformUniquePixels +[ OK ] DataTransformTest/0.TestEmptyTransformUniquePixels (0 ms) +[ RUN ] DataTransformTest/0.TestMirrorTrain +[ OK ] DataTransformTest/0.TestMirrorTrain (0 ms) +[ RUN ] DataTransformTest/0.TestCropMirrorTrain +[ OK ] DataTransformTest/0.TestCropMirrorTrain (0 ms) +[ RUN ] DataTransformTest/0.TestEmptyTransform +[ OK ] DataTransformTest/0.TestEmptyTransform (0 ms) +[ RUN ] DataTransformTest/0.TestCropMirrorTest +[ OK ] DataTransformTest/0.TestCropMirrorTest (0 ms) +[ RUN ] DataTransformTest/0.TestMeanValue +[ OK ] DataTransformTest/0.TestMeanValue (0 ms) +[ RUN ] DataTransformTest/0.TestMeanValues +[ OK ] DataTransformTest/0.TestMeanValues (0 ms) +[ RUN ] DataTransformTest/0.TestMeanFile +[ OK ] DataTransformTest/0.TestMeanFile (0 ms) +[ RUN ] DataTransformTest/0.TestCropTrain +[ OK ] DataTransformTest/0.TestCropTrain (1 ms) +[ RUN ] DataTransformTest/0.TestCropTest +[ OK ] DataTransformTest/0.TestCropTest (0 ms) +[----------] 12 tests from DataTransformTest/0 (2 ms total) + +[----------] 10 tests from EltwiseLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] EltwiseLayerTest/0.TestProd +[ OK ] EltwiseLayerTest/0.TestProd (0 ms) +[ RUN ] EltwiseLayerTest/0.TestUnstableProdGradient +[ OK ] EltwiseLayerTest/0.TestUnstableProdGradient (6 ms) +[ RUN ] EltwiseLayerTest/0.TestSumGradient +[ OK ] EltwiseLayerTest/0.TestSumGradient (7 ms) +[ RUN ] EltwiseLayerTest/0.TestSetUp +[ OK ] EltwiseLayerTest/0.TestSetUp (0 ms) +[ RUN ] EltwiseLayerTest/0.TestMax +[ OK ] EltwiseLayerTest/0.TestMax (0 ms) +[ RUN ] EltwiseLayerTest/0.TestSum +[ OK ] EltwiseLayerTest/0.TestSum (0 ms) +[ RUN ] EltwiseLayerTest/0.TestSumCoeffGradient +[ OK ] EltwiseLayerTest/0.TestSumCoeffGradient (7 ms) +[ RUN ] EltwiseLayerTest/0.TestMaxGradient +[ OK ] EltwiseLayerTest/0.TestMaxGradient (8 ms) +[ RUN ] EltwiseLayerTest/0.TestSumCoeff +[ OK ] EltwiseLayerTest/0.TestSumCoeff (0 ms) +[ RUN ] EltwiseLayerTest/0.TestStableProdGradient +[ OK ] EltwiseLayerTest/0.TestStableProdGradient (6 ms) +[----------] 10 tests from EltwiseLayerTest/0 (34 ms total) + +[----------] 12 tests from NesterovSolverTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithEverythingShare +[ OK ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithEverythingShare (119 ms) +[ RUN ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithWeightDecayMultiIter +[ OK ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithWeightDecayMultiIter (111 ms) +[ RUN ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithWeightDecay +[ OK ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithWeightDecay (22 ms) +[ RUN ] NesterovSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare +[ OK ] NesterovSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare (7 ms) +[ RUN ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateLROneHundredth +[ OK ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateLROneHundredth (22 ms) +[ RUN ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithEverything +[ OK ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithEverything (111 ms) +[ RUN ] NesterovSolverTest/0.TestSnapshotShare +[ OK ] NesterovSolverTest/0.TestSnapshotShare (45 ms) +[ RUN ] NesterovSolverTest/0.TestSnapshot +[ OK ] NesterovSolverTest/0.TestSnapshot (34 ms) +[ RUN ] NesterovSolverTest/0.TestLeastSquaresUpdateWithMomentumMultiIter +[ OK ] NesterovSolverTest/0.TestLeastSquaresUpdateWithMomentumMultiIter (110 ms) +[ RUN ] NesterovSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum +[ OK ] NesterovSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum (5 ms) +[ RUN ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithMomentum +[ OK ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithMomentum (44 ms) +[ RUN ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdate +[ OK ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdate (22 ms) +[----------] 12 tests from NesterovSolverTest/0 (652 ms total) + +[----------] 5 tests from EmbedLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] EmbedLayerTest/0.TestSetUp +[ OK ] EmbedLayerTest/0.TestSetUp (0 ms) +[ RUN ] EmbedLayerTest/0.TestGradient +[ OK ] EmbedLayerTest/0.TestGradient (17 ms) +[ RUN ] EmbedLayerTest/0.TestForwardWithBias +[ OK ] EmbedLayerTest/0.TestForwardWithBias (0 ms) +[ RUN ] EmbedLayerTest/0.TestForward +[ OK ] EmbedLayerTest/0.TestForward (0 ms) +[ RUN ] EmbedLayerTest/0.TestGradientWithBias +[ OK ] EmbedLayerTest/0.TestGradientWithBias (23 ms) +[----------] 5 tests from EmbedLayerTest/0 (41 ms total) + +[----------] 6 tests from MVNLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] MVNLayerTest/1.TestForward +[ OK ] MVNLayerTest/1.TestForward (0 ms) +[ RUN ] MVNLayerTest/1.TestGradient +[ OK ] MVNLayerTest/1.TestGradient (570 ms) +[ RUN ] MVNLayerTest/1.TestForwardAcrossChannels +[ OK ] MVNLayerTest/1.TestForwardAcrossChannels (0 ms) +[ RUN ] MVNLayerTest/1.TestGradientMeanOnly +[ OK ] MVNLayerTest/1.TestGradientMeanOnly (175 ms) +[ RUN ] MVNLayerTest/1.TestForwardMeanOnly +[ OK ] MVNLayerTest/1.TestForwardMeanOnly (0 ms) +[ RUN ] MVNLayerTest/1.TestGradientAcrossChannels +[ OK ] MVNLayerTest/1.TestGradientAcrossChannels (564 ms) +[----------] 6 tests from MVNLayerTest/1 (1309 ms total) + +[----------] 1 test from LayerFactoryTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] LayerFactoryTest/0.TestCreateLayer +[ OK ] LayerFactoryTest/0.TestCreateLayer (4 ms) +[----------] 1 test from LayerFactoryTest/0 (4 ms total) + +[----------] 5 tests from BenchmarkTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] BenchmarkTest/1.TestTimerStop +[ OK ] BenchmarkTest/1.TestTimerStop (0 ms) +[ RUN ] BenchmarkTest/1.TestTimerMilliSeconds +[ OK ] BenchmarkTest/1.TestTimerMilliSeconds (300 ms) +[ RUN ] BenchmarkTest/1.TestTimerStart +[ OK ] BenchmarkTest/1.TestTimerStart (0 ms) +[ RUN ] BenchmarkTest/1.TestTimerSeconds +[ OK ] BenchmarkTest/1.TestTimerSeconds (300 ms) +[ RUN ] BenchmarkTest/1.TestTimerConstructor +[ OK ] BenchmarkTest/1.TestTimerConstructor (0 ms) +[----------] 5 tests from BenchmarkTest/1 (600 ms total) + +[----------] 3 tests from ThresholdLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] ThresholdLayerTest/0.TestSetup +[ OK ] ThresholdLayerTest/0.TestSetup (1 ms) +[ RUN ] ThresholdLayerTest/0.Test2 +[ OK ] ThresholdLayerTest/0.Test2 (0 ms) +[ RUN ] ThresholdLayerTest/0.Test +[ OK ] ThresholdLayerTest/0.Test (0 ms) +[----------] 3 tests from ThresholdLayerTest/0 (1 ms total) + +[----------] 1 test from ProtoTest +[ RUN ] ProtoTest.TestSerialization +Printing in binary format. + +testTest +Printing in text format. +name: "test" +type: "Test" + +[ OK ] ProtoTest.TestSerialization (0 ms) +[----------] 1 test from ProtoTest (0 ms total) + +[----------] 14 tests from DataLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] DataLayerTest/0.TestReadCropTestLevelDB +[ OK ] DataLayerTest/0.TestReadCropTestLevelDB (3 ms) +[ RUN ] DataLayerTest/0.TestSkipLMDB +[ OK ] DataLayerTest/0.TestSkipLMDB (7 ms) +[ RUN ] DataLayerTest/0.TestReadCropTrainLMDB +[ OK ] DataLayerTest/0.TestReadCropTrainLMDB (2 ms) +[ RUN ] DataLayerTest/0.TestReadCropTrainSequenceSeededLMDB +[ OK ] DataLayerTest/0.TestReadCropTrainSequenceSeededLMDB (2 ms) +[ RUN ] DataLayerTest/0.TestReadLMDB +[ OK ] DataLayerTest/0.TestReadLMDB (4 ms) +[ RUN ] DataLayerTest/0.TestReshapeLMDB +[ OK ] DataLayerTest/0.TestReshapeLMDB (1 ms) +[ RUN ] DataLayerTest/0.TestSkipLevelDB +[ OK ] DataLayerTest/0.TestSkipLevelDB (21 ms) +[ RUN ] DataLayerTest/0.TestReadCropTrainSequenceSeededLevelDB +[ OK ] DataLayerTest/0.TestReadCropTrainSequenceSeededLevelDB (10 ms) +[ RUN ] DataLayerTest/0.TestReadCropTrainSequenceUnseededLevelDB +[ OK ] DataLayerTest/0.TestReadCropTrainSequenceUnseededLevelDB (8 ms) +[ RUN ] DataLayerTest/0.TestReadCropTrainSequenceUnseededLMDB +[ OK ] DataLayerTest/0.TestReadCropTrainSequenceUnseededLMDB (2 ms) +[ RUN ] DataLayerTest/0.TestReadCropTrainLevelDB +[ OK ] DataLayerTest/0.TestReadCropTrainLevelDB (2 ms) +[ RUN ] DataLayerTest/0.TestReadCropTestLMDB +[ OK ] DataLayerTest/0.TestReadCropTestLMDB (8 ms) +[ RUN ] DataLayerTest/0.TestReshapeLevelDB +[ OK ] DataLayerTest/0.TestReshapeLevelDB (3 ms) +[ RUN ] DataLayerTest/0.TestReadLevelDB +[ OK ] DataLayerTest/0.TestReadLevelDB (6 ms) +[----------] 14 tests from DataLayerTest/0 (79 ms total) + +[----------] 6 tests from FlattenLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] FlattenLayerTest/1.TestSetupWithEndAxis +[ OK ] FlattenLayerTest/1.TestSetupWithEndAxis (0 ms) +[ RUN ] FlattenLayerTest/1.TestForward +[ OK ] FlattenLayerTest/1.TestForward (0 ms) +[ RUN ] FlattenLayerTest/1.TestSetup +[ OK ] FlattenLayerTest/1.TestSetup (0 ms) +[ RUN ] FlattenLayerTest/1.TestSetupWithStartAndEndAxis +[ OK ] FlattenLayerTest/1.TestSetupWithStartAndEndAxis (0 ms) +[ RUN ] FlattenLayerTest/1.TestGradient +[ OK ] FlattenLayerTest/1.TestGradient (5 ms) +[ RUN ] FlattenLayerTest/1.TestSetupWithAxis +[ OK ] FlattenLayerTest/1.TestSetupWithAxis (0 ms) +[----------] 6 tests from FlattenLayerTest/1 (5 ms total) + +[----------] 4 tests from BlobSimpleTest/0, where TypeParam = float +[ RUN ] BlobSimpleTest/0.TestReshapeZero +[ OK ] BlobSimpleTest/0.TestReshapeZero (0 ms) +[ RUN ] BlobSimpleTest/0.TestInitialization +[ OK ] BlobSimpleTest/0.TestInitialization (0 ms) +[ RUN ] BlobSimpleTest/0.TestReshape +[ OK ] BlobSimpleTest/0.TestReshape (0 ms) +[ RUN ] BlobSimpleTest/0.TestLegacyBlobProtoShapeEquals +[ OK ] BlobSimpleTest/0.TestLegacyBlobProtoShapeEquals (0 ms) +[----------] 4 tests from BlobSimpleTest/0 (0 ms total) + +[----------] 10 tests from PowerLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] PowerLayerTest/0.TestPower +[ OK ] PowerLayerTest/0.TestPower (0 ms) +[ RUN ] PowerLayerTest/0.TestPowerZero +[ OK ] PowerLayerTest/0.TestPowerZero (0 ms) +[ RUN ] PowerLayerTest/0.TestPowerZeroGradient +[ OK ] PowerLayerTest/0.TestPowerZeroGradient (3 ms) +[ RUN ] PowerLayerTest/0.TestPowerOne +[ OK ] PowerLayerTest/0.TestPowerOne (0 ms) +[ RUN ] PowerLayerTest/0.TestPowerGradientShiftZero +[ OK ] PowerLayerTest/0.TestPowerGradientShiftZero (5 ms) +[ RUN ] PowerLayerTest/0.TestPowerOneGradient +[ OK ] PowerLayerTest/0.TestPowerOneGradient (7 ms) +[ RUN ] PowerLayerTest/0.TestPowerTwoScaleHalfGradient +[ OK ] PowerLayerTest/0.TestPowerTwoScaleHalfGradient (5 ms) +[ RUN ] PowerLayerTest/0.TestPowerGradient +[ OK ] PowerLayerTest/0.TestPowerGradient (5 ms) +[ RUN ] PowerLayerTest/0.TestPowerTwoGradient +[ OK ] PowerLayerTest/0.TestPowerTwoGradient (5 ms) +[ RUN ] PowerLayerTest/0.TestPowerTwo +[ OK ] PowerLayerTest/0.TestPowerTwo (0 ms) +[----------] 10 tests from PowerLayerTest/0 (30 ms total) + +[----------] 2 tests from BilinearFillerTest/1, where TypeParam = double +[ RUN ] BilinearFillerTest/1.TestFillOdd +[ OK ] BilinearFillerTest/1.TestFillOdd (21 ms) +[ RUN ] BilinearFillerTest/1.TestFillEven +[ OK ] BilinearFillerTest/1.TestFillEven (16 ms) +[----------] 2 tests from BilinearFillerTest/1 (38 ms total) + [----------] 12 tests from ArgMaxLayerTest/0, where TypeParam = float +[ RUN ] ArgMaxLayerTest/0.TestSetupAxis +[ OK ] ArgMaxLayerTest/0.TestSetupAxis (1 ms) [ RUN ] ArgMaxLayerTest/0.TestCPU -[ OK ] ArgMaxLayerTest/0.TestCPU (3 ms) -[ RUN ] ArgMaxLayerTest/0.TestCPUAxisTopK -[ OK ] ArgMaxLayerTest/0.TestCPUAxisTopK (57 ms) +[ OK ] ArgMaxLayerTest/0.TestCPU (6 ms) +[ RUN ] ArgMaxLayerTest/0.TestSetupAxisMaxVal +[ OK ] ArgMaxLayerTest/0.TestSetupAxisMaxVal (1 ms) +[ RUN ] ArgMaxLayerTest/0.TestCPUMaxValTopK +[ OK ] ArgMaxLayerTest/0.TestCPUMaxValTopK (2 ms) +[ RUN ] ArgMaxLayerTest/0.TestSetup +[ OK ] ArgMaxLayerTest/0.TestSetup (1 ms) [ RUN ] ArgMaxLayerTest/0.TestSetupMaxVal [ OK ] ArgMaxLayerTest/0.TestSetupMaxVal (1 ms) +[ RUN ] ArgMaxLayerTest/0.TestCPUAxisTopK +[ OK ] ArgMaxLayerTest/0.TestCPUAxisTopK (69 ms) [ RUN ] ArgMaxLayerTest/0.TestCPUTopK -[ OK ] ArgMaxLayerTest/0.TestCPUTopK (2 ms) +[ OK ] ArgMaxLayerTest/0.TestCPUTopK (6 ms) [ RUN ] ArgMaxLayerTest/0.TestCPUMaxVal -[ OK ] ArgMaxLayerTest/0.TestCPUMaxVal (1 ms) -[ RUN ] ArgMaxLayerTest/0.TestCPUMaxValTopK -[ OK ] ArgMaxLayerTest/0.TestCPUMaxValTopK (2 ms) +[ OK ] ArgMaxLayerTest/0.TestCPUMaxVal (2 ms) +[ RUN ] ArgMaxLayerTest/0.TestSetupAxisNegativeIndexing +[ OK ] ArgMaxLayerTest/0.TestSetupAxisNegativeIndexing (1 ms) [ RUN ] ArgMaxLayerTest/0.TestCPUAxis [ OK ] ArgMaxLayerTest/0.TestCPUAxis (14 ms) -[ RUN ] ArgMaxLayerTest/0.TestSetup -[ OK ] ArgMaxLayerTest/0.TestSetup (1 ms) [ RUN ] ArgMaxLayerTest/0.TestCPUAxisMaxValTopK -[ OK ] ArgMaxLayerTest/0.TestCPUAxisMaxValTopK (54 ms) -[ RUN ] ArgMaxLayerTest/0.TestSetupAxis -[ OK ] ArgMaxLayerTest/0.TestSetupAxis (1 ms) -[ RUN ] ArgMaxLayerTest/0.TestSetupAxisNegativeIndexing -[ OK ] ArgMaxLayerTest/0.TestSetupAxisNegativeIndexing (1 ms) -[ RUN ] ArgMaxLayerTest/0.TestSetupAxisMaxVal -[ OK ] ArgMaxLayerTest/0.TestSetupAxisMaxVal (1 ms) -[----------] 12 tests from ArgMaxLayerTest/0 (138 ms total) +[ OK ] ArgMaxLayerTest/0.TestCPUAxisMaxValTopK (59 ms) +[----------] 12 tests from ArgMaxLayerTest/0 (163 ms total) -[----------] 1 test from SolverFactoryTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] SolverFactoryTest/1.TestCreateSolver -[ OK ] SolverFactoryTest/1.TestCreateSolver (11 ms) -[----------] 1 test from SolverFactoryTest/1 (11 ms total) - -[----------] 12 tests from DataTransformTest/1, where TypeParam = double -[ RUN ] DataTransformTest/1.TestCropSize -[ OK ] DataTransformTest/1.TestCropSize (0 ms) -[ RUN ] DataTransformTest/1.TestCropTrain -[ OK ] DataTransformTest/1.TestCropTrain (0 ms) -[ RUN ] DataTransformTest/1.TestMeanValue -[ OK ] DataTransformTest/1.TestMeanValue (0 ms) -[ RUN ] DataTransformTest/1.TestCropMirrorTest -[ OK ] DataTransformTest/1.TestCropMirrorTest (0 ms) -[ RUN ] DataTransformTest/1.TestCropMirrorTrain -[ OK ] DataTransformTest/1.TestCropMirrorTrain (1 ms) -[ RUN ] DataTransformTest/1.TestMeanFile -[ OK ] DataTransformTest/1.TestMeanFile (0 ms) -[ RUN ] DataTransformTest/1.TestMeanValues -[ OK ] DataTransformTest/1.TestMeanValues (0 ms) -[ RUN ] DataTransformTest/1.TestCropTest -[ OK ] DataTransformTest/1.TestCropTest (0 ms) -[ RUN ] DataTransformTest/1.TestMirrorTrain -[ OK ] DataTransformTest/1.TestMirrorTrain (0 ms) -[ RUN ] DataTransformTest/1.TestMirrorTest -[ OK ] DataTransformTest/1.TestMirrorTest (1 ms) -[ RUN ] DataTransformTest/1.TestEmptyTransformUniquePixels -[ OK ] DataTransformTest/1.TestEmptyTransformUniquePixels (0 ms) -[ RUN ] DataTransformTest/1.TestEmptyTransform -[ OK ] DataTransformTest/1.TestEmptyTransform (0 ms) -[----------] 12 tests from DataTransformTest/1 (2 ms total) +[----------] 4 tests from SoftmaxWithLossLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] SoftmaxWithLossLayerTest/0.TestGradientIgnoreLabel +[ OK ] SoftmaxWithLossLayerTest/0.TestGradientIgnoreLabel (18 ms) +[ RUN ] SoftmaxWithLossLayerTest/0.TestForwardIgnoreLabel +[ OK ] SoftmaxWithLossLayerTest/0.TestForwardIgnoreLabel (0 ms) +[ RUN ] SoftmaxWithLossLayerTest/0.TestGradient +[ OK ] SoftmaxWithLossLayerTest/0.TestGradient (19 ms) +[ RUN ] SoftmaxWithLossLayerTest/0.TestGradientUnnormalized +[ OK ] SoftmaxWithLossLayerTest/0.TestGradientUnnormalized (17 ms) +[----------] 4 tests from SoftmaxWithLossLayerTest/0 (54 ms total) -[----------] 6 tests from MVNLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] MVNLayerTest/1.TestGradient -[ OK ] MVNLayerTest/1.TestGradient (550 ms) -[ RUN ] MVNLayerTest/1.TestForward -[ OK ] MVNLayerTest/1.TestForward (1 ms) -[ RUN ] MVNLayerTest/1.TestForwardMeanOnly -[ OK ] MVNLayerTest/1.TestForwardMeanOnly (0 ms) -[ RUN ] MVNLayerTest/1.TestGradientMeanOnly -[ OK ] MVNLayerTest/1.TestGradientMeanOnly (173 ms) -[ RUN ] MVNLayerTest/1.TestGradientAcrossChannels -[ OK ] MVNLayerTest/1.TestGradientAcrossChannels (541 ms) -[ RUN ] MVNLayerTest/1.TestForwardAcrossChannels -[ OK ] MVNLayerTest/1.TestForwardAcrossChannels (0 ms) -[----------] 6 tests from MVNLayerTest/1 (1265 ms total) +[----------] 4 tests from PositiveUnitballFillerTest/0, where TypeParam = float +[ RUN ] PositiveUnitballFillerTest/0.TestFill5D +[ OK ] PositiveUnitballFillerTest/0.TestFill5D (0 ms) +[ RUN ] PositiveUnitballFillerTest/0.TestFill +[ OK ] PositiveUnitballFillerTest/0.TestFill (0 ms) +[ RUN ] PositiveUnitballFillerTest/0.TestFill1D +[ OK ] PositiveUnitballFillerTest/0.TestFill1D (1 ms) +[ RUN ] PositiveUnitballFillerTest/0.TestFill2D +[ OK ] PositiveUnitballFillerTest/0.TestFill2D (0 ms) +[----------] 4 tests from PositiveUnitballFillerTest/0 (1 ms total) -[----------] 12 tests from SGDSolverTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] SGDSolverTest/1.TestSnapshot -[ OK ] SGDSolverTest/1.TestSnapshot (34 ms) -[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdateWithWeightDecayMultiIter -[ OK ] SGDSolverTest/1.TestLeastSquaresUpdateWithWeightDecayMultiIter (106 ms) -[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdateWithMomentumMultiIter -[ OK ] SGDSolverTest/1.TestLeastSquaresUpdateWithMomentumMultiIter (106 ms) -[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdateWithEverythingShare -[ OK ] SGDSolverTest/1.TestLeastSquaresUpdateWithEverythingShare (109 ms) -[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare -[ OK ] SGDSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare (7 ms) -[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdateWithMomentum -[ OK ] SGDSolverTest/1.TestLeastSquaresUpdateWithMomentum (42 ms) -[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdate -[ OK ] SGDSolverTest/1.TestLeastSquaresUpdate (21 ms) -[ RUN ] SGDSolverTest/1.TestSnapshotShare -[ OK ] SGDSolverTest/1.TestSnapshotShare (37 ms) -[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdateLROneHundredth -[ OK ] SGDSolverTest/1.TestLeastSquaresUpdateLROneHundredth (21 ms) -[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdateWithEverything -[ OK ] SGDSolverTest/1.TestLeastSquaresUpdateWithEverything (103 ms) -[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum -[ OK ] SGDSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum (4 ms) -[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdateWithWeightDecay -[ OK ] SGDSolverTest/1.TestLeastSquaresUpdateWithWeightDecay (41 ms) -[----------] 12 tests from SGDSolverTest/1 (632 ms total) +[----------] 9 tests from LSTMLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] LSTMLayerTest/1.TestLSTMUnitGradientNonZeroCont +[ OK ] LSTMLayerTest/1.TestLSTMUnitGradientNonZeroCont (77 ms) +[ RUN ] LSTMLayerTest/1.TestForward +[ OK ] LSTMLayerTest/1.TestForward (10 ms) +[ RUN ] LSTMLayerTest/1.TestGradientNonZeroContBufferSize2WithStaticInput +[ OK ] LSTMLayerTest/1.TestGradientNonZeroContBufferSize2WithStaticInput (6509 ms) +[ RUN ] LSTMLayerTest/1.TestGradientNonZeroCont +[ OK ] LSTMLayerTest/1.TestGradientNonZeroCont (561 ms) +[ RUN ] LSTMLayerTest/1.TestLSTMUnitSetUp +[ OK ] LSTMLayerTest/1.TestLSTMUnitSetUp (0 ms) +[ RUN ] LSTMLayerTest/1.TestGradient +[ OK ] LSTMLayerTest/1.TestGradient (547 ms) +[ RUN ] LSTMLayerTest/1.TestLSTMUnitGradient +[ OK ] LSTMLayerTest/1.TestLSTMUnitGradient (80 ms) +[ RUN ] LSTMLayerTest/1.TestGradientNonZeroContBufferSize2 +[ OK ] LSTMLayerTest/1.TestGradientNonZeroContBufferSize2 (1040 ms) +[ RUN ] LSTMLayerTest/1.TestSetUp +[ OK ] LSTMLayerTest/1.TestSetUp (3 ms) +[----------] 9 tests from LSTMLayerTest/1 (8828 ms total) -[----------] 12 tests from NesterovSolverTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateLROneHundredth -[ OK ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateLROneHundredth (21 ms) -[ RUN ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithEverything -[ OK ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithEverything (103 ms) -[ RUN ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithWeightDecay -[ OK ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithWeightDecay (20 ms) -[ RUN ] NesterovSolverTest/1.TestLeastSquaresUpdateWithMomentumMultiIter -[ OK ] NesterovSolverTest/1.TestLeastSquaresUpdateWithMomentumMultiIter (102 ms) -[ RUN ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithEverythingShare -[ OK ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithEverythingShare (108 ms) -[ RUN ] NesterovSolverTest/1.TestSnapshot -[ OK ] NesterovSolverTest/1.TestSnapshot (30 ms) -[ RUN ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithWeightDecayMultiIter -[ OK ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithWeightDecayMultiIter (102 ms) -[ RUN ] NesterovSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare -[ OK ] NesterovSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare (6 ms) -[ RUN ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdate -[ OK ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdate (21 ms) -[ RUN ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithMomentum -[ OK ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithMomentum (41 ms) -[ RUN ] NesterovSolverTest/1.TestSnapshotShare -[ OK ] NesterovSolverTest/1.TestSnapshotShare (35 ms) -[ RUN ] NesterovSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum -[ OK ] NesterovSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum (4 ms) -[----------] 12 tests from NesterovSolverTest/1 (593 ms total) +[----------] 2 tests from BatchReindexLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] BatchReindexLayerTest/1.TestGradient +[ OK ] BatchReindexLayerTest/1.TestGradient (215 ms) +[ RUN ] BatchReindexLayerTest/1.TestForward +[ OK ] BatchReindexLayerTest/1.TestForward (1 ms) +[----------] 2 tests from BatchReindexLayerTest/1 (216 ms total) [----------] 1 test from CPUStochasticPoolingLayerTest/0, where TypeParam = float [ RUN ] CPUStochasticPoolingLayerTest/0.TestSetup [ OK ] CPUStochasticPoolingLayerTest/0.TestSetup (0 ms) [----------] 1 test from CPUStochasticPoolingLayerTest/0 (0 ms total) -[----------] 8 tests from Im2colLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] Im2colLayerTest/1.TestRect -[ OK ] Im2colLayerTest/1.TestRect (0 ms) -[ RUN ] Im2colLayerTest/1.TestDilatedGradientForceND -[ OK ] Im2colLayerTest/1.TestDilatedGradientForceND (2568 ms) -[ RUN ] Im2colLayerTest/1.TestForward -[ OK ] Im2colLayerTest/1.TestForward (0 ms) -[ RUN ] Im2colLayerTest/1.TestGradientForceND -[ OK ] Im2colLayerTest/1.TestGradientForceND (832 ms) -[ RUN ] Im2colLayerTest/1.TestRectGradient -[ OK ] Im2colLayerTest/1.TestRectGradient (386 ms) -[ RUN ] Im2colLayerTest/1.TestSetup -[ OK ] Im2colLayerTest/1.TestSetup (0 ms) -[ RUN ] Im2colLayerTest/1.TestDilatedGradient -[ OK ] Im2colLayerTest/1.TestDilatedGradient (1378 ms) -[ RUN ] Im2colLayerTest/1.TestGradient -[ OK ] Im2colLayerTest/1.TestGradient (473 ms) -[----------] 8 tests from Im2colLayerTest/1 (5638 ms total) - -[----------] 10 tests from ConcatLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] ConcatLayerTest/0.TestForwardNum -[ OK ] ConcatLayerTest/0.TestForwardNum (0 ms) -[ RUN ] ConcatLayerTest/0.TestForwardTrivial -[ OK ] ConcatLayerTest/0.TestForwardTrivial (0 ms) -[ RUN ] ConcatLayerTest/0.TestSetupNum -[ OK ] ConcatLayerTest/0.TestSetupNum (0 ms) -[ RUN ] ConcatLayerTest/0.TestSetupChannels -[ OK ] ConcatLayerTest/0.TestSetupChannels (0 ms) -[ RUN ] ConcatLayerTest/0.TestSetupChannelsNegativeIndexing -[ OK ] ConcatLayerTest/0.TestSetupChannelsNegativeIndexing (0 ms) -[ RUN ] ConcatLayerTest/0.TestGradientChannelsBottomOneOnly -[ OK ] ConcatLayerTest/0.TestGradientChannelsBottomOneOnly (4 ms) -[ RUN ] ConcatLayerTest/0.TestGradientChannels -[ OK ] ConcatLayerTest/0.TestGradientChannels (6 ms) -[ RUN ] ConcatLayerTest/0.TestForwardChannels -[ OK ] ConcatLayerTest/0.TestForwardChannels (1 ms) -[ RUN ] ConcatLayerTest/0.TestGradientTrivial -[ OK ] ConcatLayerTest/0.TestGradientTrivial (4 ms) -[ RUN ] ConcatLayerTest/0.TestGradientNum -[ OK ] ConcatLayerTest/0.TestGradientNum (7 ms) -[----------] 10 tests from ConcatLayerTest/0 (22 ms total) - -[----------] 4 tests from PositiveUnitballFillerTest/1, where TypeParam = double -[ RUN ] PositiveUnitballFillerTest/1.TestFill1D -[ OK ] PositiveUnitballFillerTest/1.TestFill1D (0 ms) -[ RUN ] PositiveUnitballFillerTest/1.TestFill5D -[ OK ] PositiveUnitballFillerTest/1.TestFill5D (0 ms) -[ RUN ] PositiveUnitballFillerTest/1.TestFill -[ OK ] PositiveUnitballFillerTest/1.TestFill (0 ms) -[ RUN ] PositiveUnitballFillerTest/1.TestFill2D -[ OK ] PositiveUnitballFillerTest/1.TestFill2D (0 ms) -[----------] 4 tests from PositiveUnitballFillerTest/1 (0 ms total) - -[----------] 27 tests from ReductionLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] ReductionLayerTest/1.TestAbsSumCoeffAxis1Gradient -[ OK ] ReductionLayerTest/1.TestAbsSumCoeffAxis1Gradient (3 ms) -[ RUN ] ReductionLayerTest/1.TestSumGradient -[ OK ] ReductionLayerTest/1.TestSumGradient (1 ms) -[ RUN ] ReductionLayerTest/1.TestMeanCoeffGradient -[ OK ] ReductionLayerTest/1.TestMeanCoeffGradient (1 ms) -[ RUN ] ReductionLayerTest/1.TestAbsSumGradient -[ OK ] ReductionLayerTest/1.TestAbsSumGradient (1 ms) -[ RUN ] ReductionLayerTest/1.TestSetUpWithAxis2 -[ OK ] ReductionLayerTest/1.TestSetUpWithAxis2 (0 ms) -[ RUN ] ReductionLayerTest/1.TestSumOfSquaresCoeffAxis1 -[ OK ] ReductionLayerTest/1.TestSumOfSquaresCoeffAxis1 (0 ms) -[ RUN ] ReductionLayerTest/1.TestSumCoeff -[ OK ] ReductionLayerTest/1.TestSumCoeff (0 ms) -[ RUN ] ReductionLayerTest/1.TestMean -[ OK ] ReductionLayerTest/1.TestMean (0 ms) -[ RUN ] ReductionLayerTest/1.TestSumCoeffAxis1 -[ OK ] ReductionLayerTest/1.TestSumCoeffAxis1 (0 ms) -[ RUN ] ReductionLayerTest/1.TestAbsSumCoeff -[ OK ] ReductionLayerTest/1.TestAbsSumCoeff (0 ms) -[ RUN ] ReductionLayerTest/1.TestMeanCoeffGradientAxis1 -[ OK ] ReductionLayerTest/1.TestMeanCoeffGradientAxis1 (3 ms) -[ RUN ] ReductionLayerTest/1.TestSumOfSquaresCoeff -[ OK ] ReductionLayerTest/1.TestSumOfSquaresCoeff (0 ms) -[ RUN ] ReductionLayerTest/1.TestSetUp -[ OK ] ReductionLayerTest/1.TestSetUp (0 ms) -[ RUN ] ReductionLayerTest/1.TestSumOfSquares -[ OK ] ReductionLayerTest/1.TestSumOfSquares (0 ms) -[ RUN ] ReductionLayerTest/1.TestSumOfSquaresGradient -[ OK ] ReductionLayerTest/1.TestSumOfSquaresGradient (1 ms) -[ RUN ] ReductionLayerTest/1.TestSum -[ OK ] ReductionLayerTest/1.TestSum (0 ms) -[ RUN ] ReductionLayerTest/1.TestAbsSum -[ OK ] ReductionLayerTest/1.TestAbsSum (0 ms) -[ RUN ] ReductionLayerTest/1.TestSumOfSquaresCoeffGradient -[ OK ] ReductionLayerTest/1.TestSumOfSquaresCoeffGradient (1 ms) -[ RUN ] ReductionLayerTest/1.TestSumOfSquaresCoeffAxis1Gradient -[ OK ] ReductionLayerTest/1.TestSumOfSquaresCoeffAxis1Gradient (2 ms) -[ RUN ] ReductionLayerTest/1.TestAbsSumCoeffGradient -[ OK ] ReductionLayerTest/1.TestAbsSumCoeffGradient (1 ms) -[ RUN ] ReductionLayerTest/1.TestMeanGradient -[ OK ] ReductionLayerTest/1.TestMeanGradient (1 ms) -[ RUN ] ReductionLayerTest/1.TestAbsSumCoeffAxis1 -[ OK ] ReductionLayerTest/1.TestAbsSumCoeffAxis1 (0 ms) -[ RUN ] ReductionLayerTest/1.TestSumCoeffGradient -[ OK ] ReductionLayerTest/1.TestSumCoeffGradient (2 ms) -[ RUN ] ReductionLayerTest/1.TestMeanCoeffAxis1 -[ OK ] ReductionLayerTest/1.TestMeanCoeffAxis1 (0 ms) -[ RUN ] ReductionLayerTest/1.TestMeanCoeff -[ OK ] ReductionLayerTest/1.TestMeanCoeff (0 ms) -[ RUN ] ReductionLayerTest/1.TestSetUpWithAxis1 -[ OK ] ReductionLayerTest/1.TestSetUpWithAxis1 (0 ms) -[ RUN ] ReductionLayerTest/1.TestSumCoeffAxis1Gradient -[ OK ] ReductionLayerTest/1.TestSumCoeffAxis1Gradient (2 ms) -[----------] 27 tests from ReductionLayerTest/1 (19 ms total) - -[----------] 4 tests from UniformFillerTest/0, where TypeParam = float -[ RUN ] UniformFillerTest/0.TestFill -[ OK ] UniformFillerTest/0.TestFill (0 ms) -[ RUN ] UniformFillerTest/0.TestFill5D -[ OK ] UniformFillerTest/0.TestFill5D (0 ms) -[ RUN ] UniformFillerTest/0.TestFill2D -[ OK ] UniformFillerTest/0.TestFill2D (0 ms) -[ RUN ] UniformFillerTest/0.TestFill1D -[ OK ] UniformFillerTest/0.TestFill1D (0 ms) -[----------] 4 tests from UniformFillerTest/0 (0 ms total) - -[----------] 8 tests from SliceLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] SliceLayerTest/1.TestSliceAcrossChannels -[ OK ] SliceLayerTest/1.TestSliceAcrossChannels (1 ms) -[ RUN ] SliceLayerTest/1.TestSetupChannels -[ OK ] SliceLayerTest/1.TestSetupChannels (0 ms) -[ RUN ] SliceLayerTest/1.TestTrivialSlice -[ OK ] SliceLayerTest/1.TestTrivialSlice (0 ms) -[ RUN ] SliceLayerTest/1.TestSliceAcrossNum -[ OK ] SliceLayerTest/1.TestSliceAcrossNum (0 ms) -[ RUN ] SliceLayerTest/1.TestGradientAcrossNum -[ OK ] SliceLayerTest/1.TestGradientAcrossNum (58 ms) -[ RUN ] SliceLayerTest/1.TestGradientTrivial -[ OK ] SliceLayerTest/1.TestGradientTrivial (16 ms) -[ RUN ] SliceLayerTest/1.TestSetupNum -[ OK ] SliceLayerTest/1.TestSetupNum (1 ms) -[ RUN ] SliceLayerTest/1.TestGradientAcrossChannels -[ OK ] SliceLayerTest/1.TestGradientAcrossChannels (73 ms) -[----------] 8 tests from SliceLayerTest/1 (149 ms total) - -[----------] 26 tests from IOTest -[ RUN ] IOTest.TestReadImageToDatumResized -[ OK ] IOTest.TestReadImageToDatumResized (15 ms) -[ RUN ] IOTest.TestDecodeDatumToCVMatNativeGray -[ OK ] IOTest.TestDecodeDatumToCVMatNativeGray (3 ms) -[ RUN ] IOTest.TestCVMatToDatumReference -[ OK ] IOTest.TestCVMatToDatumReference (21 ms) -[ RUN ] IOTest.TestReadImageToCVMatResizedSquare -[ OK ] IOTest.TestReadImageToCVMatResizedSquare (9 ms) -[ RUN ] IOTest.TestReadFileToDatum -[ OK ] IOTest.TestReadFileToDatum (0 ms) -[ RUN ] IOTest.TestDecodeDatumToCVMat -[ OK ] IOTest.TestDecodeDatumToCVMat (9 ms) -[ RUN ] IOTest.TestDecodeDatumToCVMatNative -[ OK ] IOTest.TestDecodeDatumToCVMatNative (6 ms) -[ RUN ] IOTest.TestDecodeDatumToCVMatContent -[ OK ] IOTest.TestDecodeDatumToCVMatContent (18 ms) -[ RUN ] IOTest.TestReadImageToCVMatResizedGray -[ OK ] IOTest.TestReadImageToCVMatResizedGray (6 ms) -[ RUN ] IOTest.TestReadImageToDatum -[ OK ] IOTest.TestReadImageToDatum (7 ms) -[ RUN ] IOTest.TestReadImageToDatumReference -[ OK ] IOTest.TestReadImageToDatumReference (19 ms) -[ RUN ] IOTest.TestDecodeDatumNativeGray -[ OK ] IOTest.TestDecodeDatumNativeGray (9 ms) -[ RUN ] IOTest.TestReadImageToDatumReferenceResized -[ OK ] IOTest.TestReadImageToDatumReferenceResized (15 ms) -[ RUN ] IOTest.TestCVMatToDatumContent -[ OK ] IOTest.TestCVMatToDatumContent (17 ms) -[ RUN ] IOTest.TestReadImageToCVMatGray -[ OK ] IOTest.TestReadImageToCVMatGray (4 ms) -[ RUN ] IOTest.TestReadImageToDatumResizedSquare -[ OK ] IOTest.TestReadImageToDatumResizedSquare (8 ms) -[ RUN ] IOTest.TestReadImageToDatumContent -[ OK ] IOTest.TestReadImageToDatumContent (18 ms) -[ RUN ] IOTest.TestDecodeDatumNative -[ OK ] IOTest.TestDecodeDatumNative (21 ms) -[ RUN ] IOTest.TestReadImageToDatumContentGray -[ OK ] IOTest.TestReadImageToDatumContentGray (10 ms) -[ RUN ] IOTest.TestDecodeDatum -[ OK ] IOTest.TestDecodeDatum (22 ms) -[ RUN ] IOTest.TestReadImageToDatumResizedGray -[ OK ] IOTest.TestReadImageToDatumResizedGray (5 ms) -[ RUN ] IOTest.TestDecodeDatumToCVMatContentNative -[ OK ] IOTest.TestDecodeDatumToCVMatContentNative (18 ms) -[ RUN ] IOTest.TestReadImageToCVMat -[ OK ] IOTest.TestReadImageToCVMat (7 ms) -[ RUN ] IOTest.TestCVMatToDatum -[ OK ] IOTest.TestCVMatToDatum (7 ms) -[ RUN ] IOTest.TestReadImageToDatumGray -[ OK ] IOTest.TestReadImageToDatumGray (5 ms) -[ RUN ] IOTest.TestReadImageToCVMatResized -[ OK ] IOTest.TestReadImageToCVMatResized (7 ms) -[----------] 26 tests from IOTest (287 ms total) +[----------] 15 tests from ConvolutionLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] ConvolutionLayerTest/0.TestSetup +[ OK ] ConvolutionLayerTest/0.TestSetup (0 ms) +[ RUN ] ConvolutionLayerTest/0.TestSobelConvolution +[ OK ] ConvolutionLayerTest/0.TestSobelConvolution (0 ms) +[ RUN ] ConvolutionLayerTest/0.TestGradientGroup +[ OK ] ConvolutionLayerTest/0.TestGradientGroup (37 ms) +[ RUN ] ConvolutionLayerTest/0.TestDilatedGradient +[ OK ] ConvolutionLayerTest/0.TestDilatedGradient (34 ms) +[ RUN ] ConvolutionLayerTest/0.TestSimple3DConvolution +[ OK ] ConvolutionLayerTest/0.TestSimple3DConvolution (3 ms) +[ RUN ] ConvolutionLayerTest/0.TestGradient +[ OK ] ConvolutionLayerTest/0.TestGradient (127 ms) +[ RUN ] ConvolutionLayerTest/0.Test0DConvolution +[ OK ] ConvolutionLayerTest/0.Test0DConvolution (0 ms) +[ RUN ] ConvolutionLayerTest/0.TestDilated3DConvolution +[ OK ] ConvolutionLayerTest/0.TestDilated3DConvolution (16 ms) +[ RUN ] ConvolutionLayerTest/0.TestGradient3D +[ OK ] ConvolutionLayerTest/0.TestGradient3D (1109 ms) +[ RUN ] ConvolutionLayerTest/0.TestSimpleConvolutionGroup +[ OK ] ConvolutionLayerTest/0.TestSimpleConvolutionGroup (0 ms) +[ RUN ] ConvolutionLayerTest/0.Test1x1Convolution +[ OK ] ConvolutionLayerTest/0.Test1x1Convolution (1 ms) +[ RUN ] ConvolutionLayerTest/0.TestSimpleConvolution +[ OK ] ConvolutionLayerTest/0.TestSimpleConvolution (0 ms) +[ RUN ] ConvolutionLayerTest/0.TestDilatedConvolution +[ OK ] ConvolutionLayerTest/0.TestDilatedConvolution (3 ms) +[ RUN ] ConvolutionLayerTest/0.Test1x1Gradient +[ OK ] ConvolutionLayerTest/0.Test1x1Gradient (985 ms) +[ RUN ] ConvolutionLayerTest/0.TestNDAgainst2D +[ OK ] ConvolutionLayerTest/0.TestNDAgainst2D (682 ms) +[----------] 15 tests from ConvolutionLayerTest/0 (2997 ms total) -[----------] 2 tests from BatchReindexLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] BatchReindexLayerTest/0.TestGradient -[ OK ] BatchReindexLayerTest/0.TestGradient (194 ms) -[ RUN ] BatchReindexLayerTest/0.TestForward -[ OK ] BatchReindexLayerTest/0.TestForward (0 ms) -[----------] 2 tests from BatchReindexLayerTest/0 (194 ms total) +[----------] 5 tests from SPPLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] SPPLayerTest/1.TestSetup +[ OK ] SPPLayerTest/1.TestSetup (0 ms) +[ RUN ] SPPLayerTest/1.TestForwardBackward +[ OK ] SPPLayerTest/1.TestForwardBackward (0 ms) +[ RUN ] SPPLayerTest/1.TestGradient +[ OK ] SPPLayerTest/1.TestGradient (4040 ms) +[ RUN ] SPPLayerTest/1.TestEqualOutputDims +[ OK ] SPPLayerTest/1.TestEqualOutputDims (0 ms) +[ RUN ] SPPLayerTest/1.TestEqualOutputDims2 +[ OK ] SPPLayerTest/1.TestEqualOutputDims2 (0 ms) +[----------] 5 tests from SPPLayerTest/1 (4040 ms total) -[----------] 6 tests from XavierFillerTest/1, where TypeParam = double -[ RUN ] XavierFillerTest/1.TestFill5D -[ OK ] XavierFillerTest/1.TestFill5D (0 ms) -[ RUN ] XavierFillerTest/1.TestFill1D -[ OK ] XavierFillerTest/1.TestFill1D (0 ms) -[ RUN ] XavierFillerTest/1.TestFillFanIn -[ OK ] XavierFillerTest/1.TestFillFanIn (64 ms) -[ RUN ] XavierFillerTest/1.TestFillAverage -[ OK ] XavierFillerTest/1.TestFillAverage (64 ms) -[ RUN ] XavierFillerTest/1.TestFillFanOut -[ OK ] XavierFillerTest/1.TestFillFanOut (64 ms) -[ RUN ] XavierFillerTest/1.TestFill2D -[ OK ] XavierFillerTest/1.TestFill2D (0 ms) -[----------] 6 tests from XavierFillerTest/1 (193 ms total) +[----------] 5 tests from DBTest/0, where TypeParam = caffe::TypeLevelDB +[ RUN ] DBTest/0.TestSeekToFirst +[ OK ] DBTest/0.TestSeekToFirst (36 ms) +[ RUN ] DBTest/0.TestNext +[ OK ] DBTest/0.TestNext (26 ms) +[ RUN ] DBTest/0.TestGetDB +[ OK ] DBTest/0.TestGetDB (17 ms) +[ RUN ] DBTest/0.TestWrite +[ OK ] DBTest/0.TestWrite (27 ms) +[ RUN ] DBTest/0.TestKeyValue +[ OK ] DBTest/0.TestKeyValue (27 ms) +[----------] 5 tests from DBTest/0 (133 ms total) -[----------] 14 tests from DataLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] DataLayerTest/1.TestReadCropTestLevelDB -[ OK ] DataLayerTest/1.TestReadCropTestLevelDB (3 ms) -[ RUN ] DataLayerTest/1.TestReadCropTrainSequenceUnseededLMDB -[ OK ] DataLayerTest/1.TestReadCropTrainSequenceUnseededLMDB (1 ms) -[ RUN ] DataLayerTest/1.TestReadCropTestLMDB -[ OK ] DataLayerTest/1.TestReadCropTestLMDB (1 ms) -[ RUN ] DataLayerTest/1.TestReadCropTrainSequenceSeededLMDB -[ OK ] DataLayerTest/1.TestReadCropTrainSequenceSeededLMDB (2 ms) -[ RUN ] DataLayerTest/1.TestReadCropTrainSequenceUnseededLevelDB -[ OK ] DataLayerTest/1.TestReadCropTrainSequenceUnseededLevelDB (3 ms) -[ RUN ] DataLayerTest/1.TestReshapeLMDB -[ OK ] DataLayerTest/1.TestReshapeLMDB (1 ms) -[ RUN ] DataLayerTest/1.TestSkipLevelDB -[ OK ] DataLayerTest/1.TestSkipLevelDB (14 ms) -[ RUN ] DataLayerTest/1.TestReshapeLevelDB -[ OK ] DataLayerTest/1.TestReshapeLevelDB (2 ms) -[ RUN ] DataLayerTest/1.TestReadCropTrainSequenceSeededLevelDB -[ OK ] DataLayerTest/1.TestReadCropTrainSequenceSeededLevelDB (3 ms) -[ RUN ] DataLayerTest/1.TestSkipLMDB -[ OK ] DataLayerTest/1.TestSkipLMDB (8 ms) -[ RUN ] DataLayerTest/1.TestReadLMDB -[ OK ] DataLayerTest/1.TestReadLMDB (5 ms) -[ RUN ] DataLayerTest/1.TestReadCropTrainLMDB -[ OK ] DataLayerTest/1.TestReadCropTrainLMDB (1 ms) -[ RUN ] DataLayerTest/1.TestReadCropTrainLevelDB -[ OK ] DataLayerTest/1.TestReadCropTrainLevelDB (1 ms) -[ RUN ] DataLayerTest/1.TestReadLevelDB -[ OK ] DataLayerTest/1.TestReadLevelDB (5 ms) -[----------] 14 tests from DataLayerTest/1 (51 ms total) +[----------] 1 test from SolverFactoryTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] SolverFactoryTest/1.TestCreateSolver +[ OK ] SolverFactoryTest/1.TestCreateSolver (2 ms) +[----------] 1 test from SolverFactoryTest/1 (2 ms total) -[----------] 1 test from SolverTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] SolverTest/1.TestInitTrainTestNets -[ OK ] SolverTest/1.TestInitTrainTestNets (3 ms) -[----------] 1 test from SolverTest/1 (3 ms total) +[----------] 3 tests from SigmoidCrossEntropyLossLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] SigmoidCrossEntropyLossLayerTest/1.TestGradient +[ OK ] SigmoidCrossEntropyLossLayerTest/1.TestGradient (1 ms) +[ RUN ] SigmoidCrossEntropyLossLayerTest/1.TestIgnoreGradient +[ OK ] SigmoidCrossEntropyLossLayerTest/1.TestIgnoreGradient (0 ms) +[ RUN ] SigmoidCrossEntropyLossLayerTest/1.TestSigmoidCrossEntropyLoss +[ OK ] SigmoidCrossEntropyLossLayerTest/1.TestSigmoidCrossEntropyLoss (3 ms) +[----------] 3 tests from SigmoidCrossEntropyLossLayerTest/1 (4 ms total) -[----------] 3 tests from BlobMathTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] BlobMathTest/0.TestScaleData -[ OK ] BlobMathTest/0.TestScaleData (0 ms) -[ RUN ] BlobMathTest/0.TestAsum -[ OK ] BlobMathTest/0.TestAsum (0 ms) -[ RUN ] BlobMathTest/0.TestSumOfSquares -[ OK ] BlobMathTest/0.TestSumOfSquares (0 ms) -[----------] 3 tests from BlobMathTest/0 (0 ms total) +[----------] 4 tests from GaussianFillerTest/0, where TypeParam = float +[ RUN ] GaussianFillerTest/0.TestFill1D +[ OK ] GaussianFillerTest/0.TestFill1D (1 ms) +[ RUN ] GaussianFillerTest/0.TestFill5D +[ OK ] GaussianFillerTest/0.TestFill5D (0 ms) +[ RUN ] GaussianFillerTest/0.TestFill2D +[ OK ] GaussianFillerTest/0.TestFill2D (0 ms) +[ RUN ] GaussianFillerTest/0.TestFill +[ OK ] GaussianFillerTest/0.TestFill (0 ms) +[----------] 4 tests from GaussianFillerTest/0 (2 ms total) -[----------] 3 tests from SigmoidCrossEntropyLossLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] SigmoidCrossEntropyLossLayerTest/0.TestSigmoidCrossEntropyLoss -[ OK ] SigmoidCrossEntropyLossLayerTest/0.TestSigmoidCrossEntropyLoss (2 ms) -[ RUN ] SigmoidCrossEntropyLossLayerTest/0.TestGradient -[ OK ] SigmoidCrossEntropyLossLayerTest/0.TestGradient (2 ms) -[ RUN ] SigmoidCrossEntropyLossLayerTest/0.TestIgnoreGradient -[ OK ] SigmoidCrossEntropyLossLayerTest/0.TestIgnoreGradient (0 ms) -[----------] 3 tests from SigmoidCrossEntropyLossLayerTest/0 (4 ms total) +[----------] 1 test from MultinomialLogisticLossLayerTest/0, where TypeParam = float +[ RUN ] MultinomialLogisticLossLayerTest/0.TestGradientCPU +[ OK ] MultinomialLogisticLossLayerTest/0.TestGradientCPU (1 ms) +[----------] 1 test from MultinomialLogisticLossLayerTest/0 (1 ms total) -[----------] 11 tests from PoolingLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] PoolingLayerTest/0.TestSetup -[ OK ] PoolingLayerTest/0.TestSetup (0 ms) -[ RUN ] PoolingLayerTest/0.TestSetupPadded -[ OK ] PoolingLayerTest/0.TestSetupPadded (0 ms) -[ RUN ] PoolingLayerTest/0.TestForwardAve -[ OK ] PoolingLayerTest/0.TestForwardAve (0 ms) -[ RUN ] PoolingLayerTest/0.TestForwardMaxTopMask -[ OK ] PoolingLayerTest/0.TestForwardMaxTopMask (0 ms) -[ RUN ] PoolingLayerTest/0.TestGradientAvePadded -[ OK ] PoolingLayerTest/0.TestGradientAvePadded (1454 ms) -[ RUN ] PoolingLayerTest/0.TestForwardMax -[ OK ] PoolingLayerTest/0.TestForwardMax (0 ms) -[ RUN ] PoolingLayerTest/0.TestSetupGlobalPooling -[ OK ] PoolingLayerTest/0.TestSetupGlobalPooling (0 ms) -[ RUN ] PoolingLayerTest/0.TestGradientMaxTopMask -[ OK ] PoolingLayerTest/0.TestGradientMaxTopMask (908 ms) -[ RUN ] PoolingLayerTest/0.TestForwardMaxPadded -[ OK ] PoolingLayerTest/0.TestForwardMaxPadded (0 ms) -[ RUN ] PoolingLayerTest/0.TestGradientMax -[ OK ] PoolingLayerTest/0.TestGradientMax (1184 ms) -[ RUN ] PoolingLayerTest/0.TestGradientAve -[ OK ] PoolingLayerTest/0.TestGradientAve (356 ms) -[----------] 11 tests from PoolingLayerTest/0 (3902 ms total) +[----------] 8 tests from LRNLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] LRNLayerTest/0.TestForwardAcrossChannels +[ OK ] LRNLayerTest/0.TestForwardAcrossChannels (0 ms) +[ RUN ] LRNLayerTest/0.TestForwardAcrossChannelsLargeRegion +[ OK ] LRNLayerTest/0.TestForwardAcrossChannelsLargeRegion (1 ms) +[ RUN ] LRNLayerTest/0.TestGradientAcrossChannels +[ OK ] LRNLayerTest/0.TestGradientAcrossChannels (1216 ms) +[ RUN ] LRNLayerTest/0.TestSetupWithinChannel +[ OK ] LRNLayerTest/0.TestSetupWithinChannel (0 ms) +[ RUN ] LRNLayerTest/0.TestGradientWithinChannel +[ OK ] LRNLayerTest/0.TestGradientWithinChannel (998 ms) +[ RUN ] LRNLayerTest/0.TestForwardWithinChannel +[ OK ] LRNLayerTest/0.TestForwardWithinChannel (0 ms) +[ RUN ] LRNLayerTest/0.TestSetupAcrossChannels +[ OK ] LRNLayerTest/0.TestSetupAcrossChannels (0 ms) +[ RUN ] LRNLayerTest/0.TestGradientAcrossChannelsLargeRegion +[ OK ] LRNLayerTest/0.TestGradientAcrossChannelsLargeRegion (1564 ms) +[----------] 8 tests from LRNLayerTest/0 (3779 ms total) -[----------] 2 tests from EuclideanLossLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] EuclideanLossLayerTest/1.TestGradient -[ OK ] EuclideanLossLayerTest/1.TestGradient (0 ms) -[ RUN ] EuclideanLossLayerTest/1.TestForward -[ OK ] EuclideanLossLayerTest/1.TestForward (0 ms) -[----------] 2 tests from EuclideanLossLayerTest/1 (1 ms total) +[----------] 6 tests from MSRAFillerTest/1, where TypeParam = double +[ RUN ] MSRAFillerTest/1.TestFillFanIn +[ OK ] MSRAFillerTest/1.TestFillFanIn (135 ms) +[ RUN ] MSRAFillerTest/1.TestFill2D +[ OK ] MSRAFillerTest/1.TestFill2D (0 ms) +[ RUN ] MSRAFillerTest/1.TestFillAverage +[ OK ] MSRAFillerTest/1.TestFillAverage (136 ms) +[ RUN ] MSRAFillerTest/1.TestFill5D +[ OK ] MSRAFillerTest/1.TestFill5D (0 ms) +[ RUN ] MSRAFillerTest/1.TestFillFanOut +[ OK ] MSRAFillerTest/1.TestFillFanOut (136 ms) +[ RUN ] MSRAFillerTest/1.TestFill1D +[ OK ] MSRAFillerTest/1.TestFill1D (0 ms) +[----------] 6 tests from MSRAFillerTest/1 (407 ms total) -[----------] 3 tests from BatchNormLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] BatchNormLayerTest/0.TestForward -[ OK ] BatchNormLayerTest/0.TestForward (0 ms) -[ RUN ] BatchNormLayerTest/0.TestGradient -[ OK ] BatchNormLayerTest/0.TestGradient (265 ms) -[ RUN ] BatchNormLayerTest/0.TestForwardInplace -[ OK ] BatchNormLayerTest/0.TestForwardInplace (0 ms) -[----------] 3 tests from BatchNormLayerTest/0 (265 ms total) +[----------] 2 tests from EuclideanLossLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] EuclideanLossLayerTest/0.TestGradient +[ OK ] EuclideanLossLayerTest/0.TestGradient (1 ms) +[ RUN ] EuclideanLossLayerTest/0.TestForward +[ OK ] EuclideanLossLayerTest/0.TestForward (0 ms) +[----------] 2 tests from EuclideanLossLayerTest/0 (1 ms total) -[----------] 22 tests from ScaleLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] ScaleLayerTest/0.TestBackwardBroadcastMiddleInPlace -[ OK ] ScaleLayerTest/0.TestBackwardBroadcastMiddleInPlace (0 ms) -[ RUN ] ScaleLayerTest/0.TestGradientBroadcastMiddleWithParam -[ OK ] ScaleLayerTest/0.TestGradientBroadcastMiddleWithParam (168 ms) -[ RUN ] ScaleLayerTest/0.TestForwardBroadcastMiddleWithParamAndBias -[ OK ] ScaleLayerTest/0.TestForwardBroadcastMiddleWithParamAndBias (0 ms) -[ RUN ] ScaleLayerTest/0.TestGradientBroadcastMiddle -[ OK ] ScaleLayerTest/0.TestGradientBroadcastMiddle (168 ms) -[ RUN ] ScaleLayerTest/0.TestGradientBroadcastBegin -[ OK ] ScaleLayerTest/0.TestGradientBroadcastBegin (139 ms) -[ RUN ] ScaleLayerTest/0.TestGradientScaleAxis2 -[ OK ] ScaleLayerTest/0.TestGradientScaleAxis2 (122 ms) -[ RUN ] ScaleLayerTest/0.TestBackwardEltwiseInPlace -[ OK ] ScaleLayerTest/0.TestBackwardEltwiseInPlace (0 ms) -[ RUN ] ScaleLayerTest/0.TestForwardScale -[ OK ] ScaleLayerTest/0.TestForwardScale (0 ms) -[ RUN ] ScaleLayerTest/0.TestForwardEltwiseWithParam -[ OK ] ScaleLayerTest/0.TestForwardEltwiseWithParam (0 ms) -[ RUN ] ScaleLayerTest/0.TestForwardBroadcastMiddleInPlace -[ OK ] ScaleLayerTest/0.TestForwardBroadcastMiddleInPlace (0 ms) -[ RUN ] ScaleLayerTest/0.TestGradientEltwise -[ OK ] ScaleLayerTest/0.TestGradientEltwise (9 ms) -[ RUN ] ScaleLayerTest/0.TestGradientBroadcastEnd -[ OK ] ScaleLayerTest/0.TestGradientBroadcastEnd (318 ms) -[ RUN ] ScaleLayerTest/0.TestGradientScaleAndBias -[ OK ] ScaleLayerTest/0.TestGradientScaleAndBias (161 ms) -[ RUN ] ScaleLayerTest/0.TestForwardEltwise -[ OK ] ScaleLayerTest/0.TestForwardEltwise (0 ms) -[ RUN ] ScaleLayerTest/0.TestForwardBroadcastEnd -[ OK ] ScaleLayerTest/0.TestForwardBroadcastEnd (0 ms) -[ RUN ] ScaleLayerTest/0.TestGradientScale -[ OK ] ScaleLayerTest/0.TestGradientScale (121 ms) -[ RUN ] ScaleLayerTest/0.TestForwardBroadcastMiddle -[ OK ] ScaleLayerTest/0.TestForwardBroadcastMiddle (0 ms) -[ RUN ] ScaleLayerTest/0.TestForwardBroadcastMiddleWithParam -[ OK ] ScaleLayerTest/0.TestForwardBroadcastMiddleWithParam (0 ms) -[ RUN ] ScaleLayerTest/0.TestForwardBroadcastBegin -[ OK ] ScaleLayerTest/0.TestForwardBroadcastBegin (0 ms) -[ RUN ] ScaleLayerTest/0.TestForwardScaleAxis2 -[ OK ] ScaleLayerTest/0.TestForwardScaleAxis2 (1 ms) -[ RUN ] ScaleLayerTest/0.TestForwardEltwiseInPlace -[ OK ] ScaleLayerTest/0.TestForwardEltwiseInPlace (0 ms) -[ RUN ] ScaleLayerTest/0.TestGradientEltwiseWithParam -[ OK ] ScaleLayerTest/0.TestGradientEltwiseWithParam (544 ms) -[----------] 22 tests from ScaleLayerTest/0 (1752 ms total) +[----------] 4 tests from ContrastiveLossLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] ContrastiveLossLayerTest/0.TestGradient +[ OK ] ContrastiveLossLayerTest/0.TestGradient (352 ms) +[ RUN ] ContrastiveLossLayerTest/0.TestGradientLegacy +[ OK ] ContrastiveLossLayerTest/0.TestGradientLegacy (326 ms) +[ RUN ] ContrastiveLossLayerTest/0.TestForwardLegacy +[ OK ] ContrastiveLossLayerTest/0.TestForwardLegacy (1 ms) +[ RUN ] ContrastiveLossLayerTest/0.TestForward +[ OK ] ContrastiveLossLayerTest/0.TestForward (0 ms) +[----------] 4 tests from ContrastiveLossLayerTest/0 (679 ms total) [----------] 58 tests from NeuronLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] NeuronLayerTest/0.TestPReLUGradientChannelShared -[ OK ] NeuronLayerTest/0.TestPReLUGradientChannelShared (150 ms) +[ RUN ] NeuronLayerTest/0.TestExpLayer +[ OK ] NeuronLayerTest/0.TestExpLayer (0 ms) +[ RUN ] NeuronLayerTest/0.TestAbsGradient +[ OK ] NeuronLayerTest/0.TestAbsGradient (3 ms) +[ RUN ] NeuronLayerTest/0.TestReLU +[ OK ] NeuronLayerTest/0.TestReLU (0 ms) +[ RUN ] NeuronLayerTest/0.TestELU +[ OK ] NeuronLayerTest/0.TestELU (0 ms) +[ RUN ] NeuronLayerTest/0.TestExpLayerBase2Shift1 +[ OK ] NeuronLayerTest/0.TestExpLayerBase2Shift1 (0 ms) +[ RUN ] NeuronLayerTest/0.TestPReLUConsistencyReLU +[ OK ] NeuronLayerTest/0.TestPReLUConsistencyReLU (0 ms) +[ RUN ] NeuronLayerTest/0.TestExpLayerBase2 +[ OK ] NeuronLayerTest/0.TestExpLayerBase2 (0 ms) +[ RUN ] NeuronLayerTest/0.TestSwishAsLinear +[ OK ] NeuronLayerTest/0.TestSwishAsLinear (0 ms) +[ RUN ] NeuronLayerTest/0.TestLogLayerBase2Scale3 +[ OK ] NeuronLayerTest/0.TestLogLayerBase2Scale3 (0 ms) +[ RUN ] NeuronLayerTest/0.TestTanH +[ OK ] NeuronLayerTest/0.TestTanH (0 ms) +[ RUN ] NeuronLayerTest/0.TestSwishGradient +[ OK ] NeuronLayerTest/0.TestSwishGradient (8 ms) +[ RUN ] NeuronLayerTest/0.TestDropoutTestPhase +[ OK ] NeuronLayerTest/0.TestDropoutTestPhase (0 ms) +[ RUN ] NeuronLayerTest/0.TestReLUWithNegativeSlope +[ OK ] NeuronLayerTest/0.TestReLUWithNegativeSlope (0 ms) +[ RUN ] NeuronLayerTest/0.TestExpGradientBase2Shift1Scale3 +[ OK ] NeuronLayerTest/0.TestExpGradientBase2Shift1Scale3 (4 ms) [ RUN ] NeuronLayerTest/0.TestReLUGradient [ OK ] NeuronLayerTest/0.TestReLUGradient (3 ms) +[ RUN ] NeuronLayerTest/0.TestExpGradientBase2Shift1 +[ OK ] NeuronLayerTest/0.TestExpGradientBase2Shift1 (3 ms) +[ RUN ] NeuronLayerTest/0.TestELUGradient +[ OK ] NeuronLayerTest/0.TestELUGradient (3 ms) +[ RUN ] NeuronLayerTest/0.TestTanHGradient +[ OK ] NeuronLayerTest/0.TestTanHGradient (7 ms) +[ RUN ] NeuronLayerTest/0.TestExpLayerWithShift +[ OK ] NeuronLayerTest/0.TestExpLayerWithShift (0 ms) [ RUN ] NeuronLayerTest/0.TestDropoutThreeQuarters [ OK ] NeuronLayerTest/0.TestDropoutThreeQuarters (0 ms) -[ RUN ] NeuronLayerTest/0.TestPReLUForward -[ OK ] NeuronLayerTest/0.TestPReLUForward (0 ms) -[ RUN ] NeuronLayerTest/0.TestLogGradientBase2 -[ OK ] NeuronLayerTest/0.TestLogGradientBase2 (5 ms) -[ RUN ] NeuronLayerTest/0.TestReLUWithNegativeSlope -[ OK ] NeuronLayerTest/0.TestReLUWithNegativeSlope (0 ms) -[ RUN ] NeuronLayerTest/0.TestELUGradient -[ OK ] NeuronLayerTest/0.TestELUGradient (4 ms) -[ RUN ] NeuronLayerTest/0.TestSwishAsLinearGradient -[ OK ] NeuronLayerTest/0.TestSwishAsLinearGradient (3 ms) +[ RUN ] NeuronLayerTest/0.TestLogLayerBase2Shift1 +[ OK ] NeuronLayerTest/0.TestLogLayerBase2Shift1 (0 ms) +[ RUN ] NeuronLayerTest/0.TestELUasReLU +[ OK ] NeuronLayerTest/0.TestELUasReLU (0 ms) +[ RUN ] NeuronLayerTest/0.TestExpLayerBase2Shift1Scale3 +[ OK ] NeuronLayerTest/0.TestExpLayerBase2Shift1Scale3 (0 ms) +[ RUN ] NeuronLayerTest/0.TestLogGradientBase2Shift1Scale3 +[ OK ] NeuronLayerTest/0.TestLogGradientBase2Shift1Scale3 (5 ms) +[ RUN ] NeuronLayerTest/0.TestExpLayerBase2Scale3 +[ OK ] NeuronLayerTest/0.TestExpLayerBase2Scale3 (0 ms) +[ RUN ] NeuronLayerTest/0.TestBNLL +[ OK ] NeuronLayerTest/0.TestBNLL (0 ms) +[ RUN ] NeuronLayerTest/0.TestSigmoid +[ OK ] NeuronLayerTest/0.TestSigmoid (0 ms) [ RUN ] NeuronLayerTest/0.TestReLUGradientWithNegativeSlope [ OK ] NeuronLayerTest/0.TestReLUGradientWithNegativeSlope (3 ms) +[ RUN ] NeuronLayerTest/0.TestAbsVal +[ OK ] NeuronLayerTest/0.TestAbsVal (0 ms) +[ RUN ] NeuronLayerTest/0.TestSwishAsLinearGradient +[ OK ] NeuronLayerTest/0.TestSwishAsLinearGradient (3 ms) [ RUN ] NeuronLayerTest/0.TestLogLayerBase2Shift1Scale3 [ OK ] NeuronLayerTest/0.TestLogLayerBase2Shift1Scale3 (0 ms) -[ RUN ] NeuronLayerTest/0.TestExpGradientBase2 -[ OK ] NeuronLayerTest/0.TestExpGradientBase2 (4 ms) -[ RUN ] NeuronLayerTest/0.TestTanHGradient -[ OK ] NeuronLayerTest/0.TestTanHGradient (7 ms) -[ RUN ] NeuronLayerTest/0.TestPReLUGradient -[ OK ] NeuronLayerTest/0.TestPReLUGradient (151 ms) -[ RUN ] NeuronLayerTest/0.TestLogLayerBase2 -[ OK ] NeuronLayerTest/0.TestLogLayerBase2 (0 ms) -[ RUN ] NeuronLayerTest/0.TestReLU -[ OK ] NeuronLayerTest/0.TestReLU (0 ms) -[ RUN ] NeuronLayerTest/0.TestExpGradient -[ OK ] NeuronLayerTest/0.TestExpGradient (4 ms) -[ RUN ] NeuronLayerTest/0.TestSigmoidGradient -[ OK ] NeuronLayerTest/0.TestSigmoidGradient (6 ms) -[ RUN ] NeuronLayerTest/0.TestSwishWithBetaGradient -[ OK ] NeuronLayerTest/0.TestSwishWithBetaGradient (7 ms) +[ RUN ] NeuronLayerTest/0.TestPReLUGradientChannelShared +[ OK ] NeuronLayerTest/0.TestPReLUGradientChannelShared (146 ms) +[ RUN ] NeuronLayerTest/0.TestClip +[ OK ] NeuronLayerTest/0.TestClip (0 ms) [ RUN ] NeuronLayerTest/0.TestLogLayer [ OK ] NeuronLayerTest/0.TestLogLayer (0 ms) -[ RUN ] NeuronLayerTest/0.TestExpGradientWithShift -[ OK ] NeuronLayerTest/0.TestExpGradientWithShift (4 ms) +[ RUN ] NeuronLayerTest/0.TestLogGradient +[ OK ] NeuronLayerTest/0.TestLogGradient (5 ms) [ RUN ] NeuronLayerTest/0.TestClipGradient -[ OK ] NeuronLayerTest/0.TestClipGradient (3 ms) -[ RUN ] NeuronLayerTest/0.TestLogGradientBase2Shift1Scale3 -[ OK ] NeuronLayerTest/0.TestLogGradientBase2Shift1Scale3 (5 ms) -[ RUN ] NeuronLayerTest/0.TestClip -[ OK ] NeuronLayerTest/0.TestClip (0 ms) -[ RUN ] NeuronLayerTest/0.TestExpGradientBase2Scale3 -[ OK ] NeuronLayerTest/0.TestExpGradientBase2Scale3 (4 ms) -[ RUN ] NeuronLayerTest/0.TestELUasReLUGradient -[ OK ] NeuronLayerTest/0.TestELUasReLUGradient (4 ms) +[ OK ] NeuronLayerTest/0.TestClipGradient (2 ms) +[ RUN ] NeuronLayerTest/0.TestLogGradientBase2Scale3 +[ OK ] NeuronLayerTest/0.TestLogGradientBase2Scale3 (5 ms) +[ RUN ] NeuronLayerTest/0.TestExpGradientBase2 +[ OK ] NeuronLayerTest/0.TestExpGradientBase2 (4 ms) [ RUN ] NeuronLayerTest/0.TestBNLLGradient [ OK ] NeuronLayerTest/0.TestBNLLGradient (6 ms) -[ RUN ] NeuronLayerTest/0.TestSigmoid -[ OK ] NeuronLayerTest/0.TestSigmoid (1 ms) -[ RUN ] NeuronLayerTest/0.TestExpLayerWithShift -[ OK ] NeuronLayerTest/0.TestExpLayerWithShift (0 ms) -[ RUN ] NeuronLayerTest/0.TestDropoutTestPhase -[ OK ] NeuronLayerTest/0.TestDropoutTestPhase (0 ms) -[ RUN ] NeuronLayerTest/0.TestLogGradientBase2Shift1 -[ OK ] NeuronLayerTest/0.TestLogGradientBase2Shift1 (5 ms) -[ RUN ] NeuronLayerTest/0.TestExpGradientBase2Shift1 -[ OK ] NeuronLayerTest/0.TestExpGradientBase2Shift1 (4 ms) -[ RUN ] NeuronLayerTest/0.TestExpGradientBase2Shift1Scale3 -[ OK ] NeuronLayerTest/0.TestExpGradientBase2Shift1Scale3 (4 ms) -[ RUN ] NeuronLayerTest/0.TestLogGradient -[ OK ] NeuronLayerTest/0.TestLogGradient (5 ms) [ RUN ] NeuronLayerTest/0.TestPReLUParam [ OK ] NeuronLayerTest/0.TestPReLUParam (0 ms) -[ RUN ] NeuronLayerTest/0.TestExpLayer -[ OK ] NeuronLayerTest/0.TestExpLayer (0 ms) -[ RUN ] NeuronLayerTest/0.TestLogLayerBase2Scale3 -[ OK ] NeuronLayerTest/0.TestLogLayerBase2Scale3 (0 ms) -[ RUN ] NeuronLayerTest/0.TestLogGradientBase2Scale3 -[ OK ] NeuronLayerTest/0.TestLogGradientBase2Scale3 (5 ms) -[ RUN ] NeuronLayerTest/0.TestLogLayerBase2Shift1 -[ OK ] NeuronLayerTest/0.TestLogLayerBase2Shift1 (0 ms) -[ RUN ] NeuronLayerTest/0.TestExpLayerBase2 -[ OK ] NeuronLayerTest/0.TestExpLayerBase2 (0 ms) -[ RUN ] NeuronLayerTest/0.TestAbsVal -[ OK ] NeuronLayerTest/0.TestAbsVal (0 ms) -[ RUN ] NeuronLayerTest/0.TestDropoutGradient -[ OK ] NeuronLayerTest/0.TestDropoutGradient (4 ms) [ RUN ] NeuronLayerTest/0.TestSwish [ OK ] NeuronLayerTest/0.TestSwish (0 ms) +[ RUN ] NeuronLayerTest/0.TestExpGradientWithShift +[ OK ] NeuronLayerTest/0.TestExpGradientWithShift (4 ms) +[ RUN ] NeuronLayerTest/0.TestDropoutGradientTest +[ OK ] NeuronLayerTest/0.TestDropoutGradientTest (3 ms) +[ RUN ] NeuronLayerTest/0.TestExpGradient +[ OK ] NeuronLayerTest/0.TestExpGradient (4 ms) +[ RUN ] NeuronLayerTest/0.TestLogLayerBase2 +[ OK ] NeuronLayerTest/0.TestLogLayerBase2 (0 ms) [ RUN ] NeuronLayerTest/0.TestSwishWithBeta [ OK ] NeuronLayerTest/0.TestSwishWithBeta (0 ms) +[ RUN ] NeuronLayerTest/0.TestPReLUForward +[ OK ] NeuronLayerTest/0.TestPReLUForward (0 ms) +[ RUN ] NeuronLayerTest/0.TestExpGradientBase2Scale3 +[ OK ] NeuronLayerTest/0.TestExpGradientBase2Scale3 (4 ms) +[ RUN ] NeuronLayerTest/0.TestPReLUGradient +[ OK ] NeuronLayerTest/0.TestPReLUGradient (149 ms) +[ RUN ] NeuronLayerTest/0.TestPReLUForwardChannelShared +[ OK ] NeuronLayerTest/0.TestPReLUForwardChannelShared (0 ms) +[ RUN ] NeuronLayerTest/0.TestLogGradientBase2 +[ OK ] NeuronLayerTest/0.TestLogGradientBase2 (4 ms) +[ RUN ] NeuronLayerTest/0.TestDropoutGradient +[ OK ] NeuronLayerTest/0.TestDropoutGradient (4 ms) +[ RUN ] NeuronLayerTest/0.TestSwishWithBetaGradient +[ OK ] NeuronLayerTest/0.TestSwishWithBetaGradient (7 ms) +[ RUN ] NeuronLayerTest/0.TestSigmoidGradient +[ OK ] NeuronLayerTest/0.TestSigmoidGradient (7 ms) [ RUN ] NeuronLayerTest/0.TestDropoutHalf [ OK ] NeuronLayerTest/0.TestDropoutHalf (0 ms) -[ RUN ] NeuronLayerTest/0.TestPReLUConsistencyReLU -[ OK ] NeuronLayerTest/0.TestPReLUConsistencyReLU (0 ms) -[ RUN ] NeuronLayerTest/0.TestExpLayerBase2Shift1 -[ OK ] NeuronLayerTest/0.TestExpLayerBase2Shift1 (0 ms) -[ RUN ] NeuronLayerTest/0.TestSwishGradient -[ OK ] NeuronLayerTest/0.TestSwishGradient (8 ms) -[ RUN ] NeuronLayerTest/0.TestBNLL -[ OK ] NeuronLayerTest/0.TestBNLL (0 ms) -[ RUN ] NeuronLayerTest/0.TestAbsGradient -[ OK ] NeuronLayerTest/0.TestAbsGradient (3 ms) -[ RUN ] NeuronLayerTest/0.TestELU -[ OK ] NeuronLayerTest/0.TestELU (0 ms) -[ RUN ] NeuronLayerTest/0.TestExpLayerBase2Scale3 -[ OK ] NeuronLayerTest/0.TestExpLayerBase2Scale3 (0 ms) +[ RUN ] NeuronLayerTest/0.TestLogGradientBase2Shift1 +[ OK ] NeuronLayerTest/0.TestLogGradientBase2Shift1 (5 ms) [ RUN ] NeuronLayerTest/0.TestPReLUInPlace [ OK ] NeuronLayerTest/0.TestPReLUInPlace (0 ms) -[ RUN ] NeuronLayerTest/0.TestELUasReLU -[ OK ] NeuronLayerTest/0.TestELUasReLU (0 ms) -[ RUN ] NeuronLayerTest/0.TestSwishAsLinear -[ OK ] NeuronLayerTest/0.TestSwishAsLinear (0 ms) -[ RUN ] NeuronLayerTest/0.TestTanH -[ OK ] NeuronLayerTest/0.TestTanH (1 ms) -[ RUN ] NeuronLayerTest/0.TestDropoutGradientTest -[ OK ] NeuronLayerTest/0.TestDropoutGradientTest (2 ms) -[ RUN ] NeuronLayerTest/0.TestPReLUForwardChannelShared -[ OK ] NeuronLayerTest/0.TestPReLUForwardChannelShared (0 ms) -[ RUN ] NeuronLayerTest/0.TestExpLayerBase2Shift1Scale3 -[ OK ] NeuronLayerTest/0.TestExpLayerBase2Shift1Scale3 (0 ms) -[----------] 58 tests from NeuronLayerTest/0 (416 ms total) +[ RUN ] NeuronLayerTest/0.TestELUasReLUGradient +[ OK ] NeuronLayerTest/0.TestELUasReLUGradient (4 ms) +[----------] 58 tests from NeuronLayerTest/0 (408 ms total) -[----------] 4 tests from BlobSimpleTest/1, where TypeParam = double -[ RUN ] BlobSimpleTest/1.TestInitialization -[ OK ] BlobSimpleTest/1.TestInitialization (0 ms) -[ RUN ] BlobSimpleTest/1.TestReshape -[ OK ] BlobSimpleTest/1.TestReshape (0 ms) -[ RUN ] BlobSimpleTest/1.TestReshapeZero -[ OK ] BlobSimpleTest/1.TestReshapeZero (0 ms) -[ RUN ] BlobSimpleTest/1.TestLegacyBlobProtoShapeEquals -[ OK ] BlobSimpleTest/1.TestLegacyBlobProtoShapeEquals (1 ms) -[----------] 4 tests from BlobSimpleTest/1 (1 ms total) +[----------] 1 test from LayerFactoryTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] LayerFactoryTest/1.TestCreateLayer +[ OK ] LayerFactoryTest/1.TestCreateLayer (2 ms) +[----------] 1 test from LayerFactoryTest/1 (2 ms total) -[----------] 27 tests from ReductionLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] ReductionLayerTest/0.TestSumOfSquaresCoeffAxis1Gradient -[ OK ] ReductionLayerTest/0.TestSumOfSquaresCoeffAxis1Gradient (2 ms) -[ RUN ] ReductionLayerTest/0.TestMeanGradient -[ OK ] ReductionLayerTest/0.TestMeanGradient (1 ms) -[ RUN ] ReductionLayerTest/0.TestSum -[ OK ] ReductionLayerTest/0.TestSum (0 ms) -[ RUN ] ReductionLayerTest/0.TestSumOfSquaresGradient -[ OK ] ReductionLayerTest/0.TestSumOfSquaresGradient (1 ms) -[ RUN ] ReductionLayerTest/0.TestSetUpWithAxis2 -[ OK ] ReductionLayerTest/0.TestSetUpWithAxis2 (0 ms) -[ RUN ] ReductionLayerTest/0.TestSumCoeffAxis1Gradient -[ OK ] ReductionLayerTest/0.TestSumCoeffAxis1Gradient (2 ms) -[ RUN ] ReductionLayerTest/0.TestAbsSumCoeffAxis1Gradient -[ OK ] ReductionLayerTest/0.TestAbsSumCoeffAxis1Gradient (2 ms) -[ RUN ] ReductionLayerTest/0.TestMeanCoeff -[ OK ] ReductionLayerTest/0.TestMeanCoeff (0 ms) -[ RUN ] ReductionLayerTest/0.TestAbsSumGradient -[ OK ] ReductionLayerTest/0.TestAbsSumGradient (1 ms) -[ RUN ] ReductionLayerTest/0.TestMeanCoeffGradientAxis1 -[ OK ] ReductionLayerTest/0.TestMeanCoeffGradientAxis1 (3 ms) -[ RUN ] ReductionLayerTest/0.TestMean -[ OK ] ReductionLayerTest/0.TestMean (0 ms) -[ RUN ] ReductionLayerTest/0.TestAbsSumCoeff -[ OK ] ReductionLayerTest/0.TestAbsSumCoeff (0 ms) -[ RUN ] ReductionLayerTest/0.TestSumCoeff -[ OK ] ReductionLayerTest/0.TestSumCoeff (0 ms) -[ RUN ] ReductionLayerTest/0.TestMeanCoeffAxis1 -[ OK ] ReductionLayerTest/0.TestMeanCoeffAxis1 (0 ms) -[ RUN ] ReductionLayerTest/0.TestAbsSumCoeffGradient -[ OK ] ReductionLayerTest/0.TestAbsSumCoeffGradient (1 ms) -[ RUN ] ReductionLayerTest/0.TestSumCoeffAxis1 -[ OK ] ReductionLayerTest/0.TestSumCoeffAxis1 (0 ms) -[ RUN ] ReductionLayerTest/0.TestAbsSum -[ OK ] ReductionLayerTest/0.TestAbsSum (0 ms) -[ RUN ] ReductionLayerTest/0.TestAbsSumCoeffAxis1 -[ OK ] ReductionLayerTest/0.TestAbsSumCoeffAxis1 (0 ms) -[ RUN ] ReductionLayerTest/0.TestSumOfSquares -[ OK ] ReductionLayerTest/0.TestSumOfSquares (0 ms) -[ RUN ] ReductionLayerTest/0.TestSumOfSquaresCoeffAxis1 -[ OK ] ReductionLayerTest/0.TestSumOfSquaresCoeffAxis1 (0 ms) -[ RUN ] ReductionLayerTest/0.TestSumOfSquaresCoeff -[ OK ] ReductionLayerTest/0.TestSumOfSquaresCoeff (0 ms) -[ RUN ] ReductionLayerTest/0.TestSetUpWithAxis1 -[ OK ] ReductionLayerTest/0.TestSetUpWithAxis1 (0 ms) -[ RUN ] ReductionLayerTest/0.TestSetUp -[ OK ] ReductionLayerTest/0.TestSetUp (0 ms) -[ RUN ] ReductionLayerTest/0.TestSumOfSquaresCoeffGradient -[ OK ] ReductionLayerTest/0.TestSumOfSquaresCoeffGradient (1 ms) -[ RUN ] ReductionLayerTest/0.TestSumCoeffGradient -[ OK ] ReductionLayerTest/0.TestSumCoeffGradient (1 ms) -[ RUN ] ReductionLayerTest/0.TestMeanCoeffGradient -[ OK ] ReductionLayerTest/0.TestMeanCoeffGradient (1 ms) -[ RUN ] ReductionLayerTest/0.TestSumGradient -[ OK ] ReductionLayerTest/0.TestSumGradient (2 ms) -[----------] 27 tests from ReductionLayerTest/0 (18 ms total) +[----------] 6 tests from MSRAFillerTest/0, where TypeParam = float +[ RUN ] MSRAFillerTest/0.TestFill2D +[ OK ] MSRAFillerTest/0.TestFill2D (0 ms) +[ RUN ] MSRAFillerTest/0.TestFillFanIn +[ OK ] MSRAFillerTest/0.TestFillFanIn (96 ms) +[ RUN ] MSRAFillerTest/0.TestFill1D +[ OK ] MSRAFillerTest/0.TestFill1D (0 ms) +[ RUN ] MSRAFillerTest/0.TestFill5D +[ OK ] MSRAFillerTest/0.TestFill5D (0 ms) +[ RUN ] MSRAFillerTest/0.TestFillFanOut +[ OK ] MSRAFillerTest/0.TestFillFanOut (95 ms) +[ RUN ] MSRAFillerTest/0.TestFillAverage +[ OK ] MSRAFillerTest/0.TestFillAverage (96 ms) +[----------] 6 tests from MSRAFillerTest/0 (287 ms total) -[----------] 2 tests from BilinearFillerTest/1, where TypeParam = double -[ RUN ] BilinearFillerTest/1.TestFillEven -[ OK ] BilinearFillerTest/1.TestFillEven (12 ms) -[ RUN ] BilinearFillerTest/1.TestFillOdd -[ OK ] BilinearFillerTest/1.TestFillOdd (17 ms) -[----------] 2 tests from BilinearFillerTest/1 (29 ms total) +[----------] 1 test from SolverTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] SolverTest/1.TestInitTrainTestNets +[ OK ] SolverTest/1.TestInitTrainTestNets (3 ms) +[----------] 1 test from SolverTest/1 (3 ms total) -[----------] 15 tests from ConvolutionLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] ConvolutionLayerTest/1.TestSobelConvolution -[ OK ] ConvolutionLayerTest/1.TestSobelConvolution (0 ms) -[ RUN ] ConvolutionLayerTest/1.TestGradient -[ OK ] ConvolutionLayerTest/1.TestGradient (123 ms) -[ RUN ] ConvolutionLayerTest/1.TestNDAgainst2D -[ OK ] ConvolutionLayerTest/1.TestNDAgainst2D (730 ms) -[ RUN ] ConvolutionLayerTest/1.TestSimpleConvolutionGroup -[ OK ] ConvolutionLayerTest/1.TestSimpleConvolutionGroup (0 ms) -[ RUN ] ConvolutionLayerTest/1.TestGradient3D -[ OK ] ConvolutionLayerTest/1.TestGradient3D (1078 ms) -[ RUN ] ConvolutionLayerTest/1.Test1x1Gradient -[ OK ] ConvolutionLayerTest/1.Test1x1Gradient (988 ms) -[ RUN ] ConvolutionLayerTest/1.TestDilated3DConvolution -[ OK ] ConvolutionLayerTest/1.TestDilated3DConvolution (16 ms) -[ RUN ] ConvolutionLayerTest/1.TestSetup -[ OK ] ConvolutionLayerTest/1.TestSetup (0 ms) -[ RUN ] ConvolutionLayerTest/1.TestSimpleConvolution -[ OK ] ConvolutionLayerTest/1.TestSimpleConvolution (1 ms) -[ RUN ] ConvolutionLayerTest/1.Test1x1Convolution -[ OK ] ConvolutionLayerTest/1.Test1x1Convolution (0 ms) -[ RUN ] ConvolutionLayerTest/1.Test0DConvolution -[ OK ] ConvolutionLayerTest/1.Test0DConvolution (0 ms) -[ RUN ] ConvolutionLayerTest/1.TestSimple3DConvolution -[ OK ] ConvolutionLayerTest/1.TestSimple3DConvolution (3 ms) -[ RUN ] ConvolutionLayerTest/1.TestDilatedGradient -[ OK ] ConvolutionLayerTest/1.TestDilatedGradient (32 ms) -[ RUN ] ConvolutionLayerTest/1.TestGradientGroup -[ OK ] ConvolutionLayerTest/1.TestGradientGroup (34 ms) -[ RUN ] ConvolutionLayerTest/1.TestDilatedConvolution -[ OK ] ConvolutionLayerTest/1.TestDilatedConvolution (3 ms) -[----------] 15 tests from ConvolutionLayerTest/1 (3008 ms total) +[----------] 4 tests from PositiveUnitballFillerTest/1, where TypeParam = double +[ RUN ] PositiveUnitballFillerTest/1.TestFill +[ OK ] PositiveUnitballFillerTest/1.TestFill (0 ms) +[ RUN ] PositiveUnitballFillerTest/1.TestFill2D +[ OK ] PositiveUnitballFillerTest/1.TestFill2D (0 ms) +[ RUN ] PositiveUnitballFillerTest/1.TestFill5D +[ OK ] PositiveUnitballFillerTest/1.TestFill5D (0 ms) +[ RUN ] PositiveUnitballFillerTest/1.TestFill1D +[ OK ] PositiveUnitballFillerTest/1.TestFill1D (0 ms) +[----------] 4 tests from PositiveUnitballFillerTest/1 (0 ms total) -[----------] 11 tests from CropLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] CropLayerTest/0.TestSetupShapeDefault -[ OK ] CropLayerTest/0.TestSetupShapeDefault (0 ms) -[ RUN ] CropLayerTest/0.TestCropAllGradient -[ OK ] CropLayerTest/0.TestCropAllGradient (433 ms) -[ RUN ] CropLayerTest/0.TestDimensionsCheck -[ OK ] CropLayerTest/0.TestDimensionsCheck (0 ms) -[ RUN ] CropLayerTest/0.TestSetupShapeNegativeIndexing -[ OK ] CropLayerTest/0.TestSetupShapeNegativeIndexing (0 ms) -[ RUN ] CropLayerTest/0.TestCropAllOffset -[ OK ] CropLayerTest/0.TestCropAllOffset (0 ms) -[ RUN ] CropLayerTest/0.TestCrop5D -[ OK ] CropLayerTest/0.TestCrop5D (0 ms) -[ RUN ] CropLayerTest/0.TestSetupShapeAll -[ OK ] CropLayerTest/0.TestSetupShapeAll (0 ms) -[ RUN ] CropLayerTest/0.TestCropHW -[ OK ] CropLayerTest/0.TestCropHW (1 ms) -[ RUN ] CropLayerTest/0.TestCropHWGradient -[ OK ] CropLayerTest/0.TestCropHWGradient (716 ms) -[ RUN ] CropLayerTest/0.TestCrop5DGradient -[ OK ] CropLayerTest/0.TestCrop5DGradient (2740 ms) -[ RUN ] CropLayerTest/0.TestCropAll -[ OK ] CropLayerTest/0.TestCropAll (1 ms) -[----------] 11 tests from CropLayerTest/0 (3891 ms total) +[----------] 2 tests from SoftmaxLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] SoftmaxLayerTest/1.TestGradient +[ OK ] SoftmaxLayerTest/1.TestGradient (348 ms) +[ RUN ] SoftmaxLayerTest/1.TestForward +[ OK ] SoftmaxLayerTest/1.TestForward (0 ms) +[----------] 2 tests from SoftmaxLayerTest/1 (348 ms total) -[----------] 4 tests from UniformFillerTest/1, where TypeParam = double -[ RUN ] UniformFillerTest/1.TestFill1D -[ OK ] UniformFillerTest/1.TestFill1D (0 ms) -[ RUN ] UniformFillerTest/1.TestFill5D -[ OK ] UniformFillerTest/1.TestFill5D (0 ms) -[ RUN ] UniformFillerTest/1.TestFill2D -[ OK ] UniformFillerTest/1.TestFill2D (0 ms) -[ RUN ] UniformFillerTest/1.TestFill -[ OK ] UniformFillerTest/1.TestFill (0 ms) -[----------] 4 tests from UniformFillerTest/1 (0 ms total) +[----------] 1 test from HDF5OutputLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] HDF5OutputLayerTest/0.TestForward +[ OK ] HDF5OutputLayerTest/0.TestForward (5 ms) +[----------] 1 test from HDF5OutputLayerTest/0 (5 ms total) -[----------] 2 tests from InfogainLossLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] InfogainLossLayerTest/1.TestGradient -[ OK ] InfogainLossLayerTest/1.TestGradient (3 ms) -[ RUN ] InfogainLossLayerTest/1.TestInfogainLoss -[ OK ] InfogainLossLayerTest/1.TestInfogainLoss (0 ms) -[----------] 2 tests from InfogainLossLayerTest/1 (3 ms total) +[----------] 4 tests from NetUpgradeTest +[ RUN ] NetUpgradeTest.TestImageNet +[ OK ] NetUpgradeTest.TestImageNet (6 ms) +[ RUN ] NetUpgradeTest.TestUpgradeV1LayerType +[ OK ] NetUpgradeTest.TestUpgradeV1LayerType (2 ms) +[ RUN ] NetUpgradeTest.TestAllParams +[ OK ] NetUpgradeTest.TestAllParams (2 ms) +[ RUN ] NetUpgradeTest.TestSimple +[ OK ] NetUpgradeTest.TestSimple (1 ms) +[----------] 4 tests from NetUpgradeTest (12 ms total) -[----------] 5 tests from BenchmarkTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] BenchmarkTest/0.TestTimerStop -[ OK ] BenchmarkTest/0.TestTimerStop (0 ms) -[ RUN ] BenchmarkTest/0.TestTimerMilliSeconds -[ OK ] BenchmarkTest/0.TestTimerMilliSeconds (301 ms) -[ RUN ] BenchmarkTest/0.TestTimerSeconds -[ OK ] BenchmarkTest/0.TestTimerSeconds (300 ms) -[ RUN ] BenchmarkTest/0.TestTimerStart -[ OK ] BenchmarkTest/0.TestTimerStart (0 ms) -[ RUN ] BenchmarkTest/0.TestTimerConstructor -[ OK ] BenchmarkTest/0.TestTimerConstructor (0 ms) -[----------] 5 tests from BenchmarkTest/0 (601 ms total) +[----------] 20 tests from BiasLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] BiasLayerTest/0.TestForwardEltwise +[ OK ] BiasLayerTest/0.TestForwardEltwise (0 ms) +[ RUN ] BiasLayerTest/0.TestForwardBroadcastBegin +[ OK ] BiasLayerTest/0.TestForwardBroadcastBegin (0 ms) +[ RUN ] BiasLayerTest/0.TestBackwardEltwiseInPlace +[ OK ] BiasLayerTest/0.TestBackwardEltwiseInPlace (0 ms) +[ RUN ] BiasLayerTest/0.TestGradientEltwise +[ OK ] BiasLayerTest/0.TestGradientEltwise (5 ms) +[ RUN ] BiasLayerTest/0.TestGradientBroadcastMiddleWithParam +[ OK ] BiasLayerTest/0.TestGradientBroadcastMiddleWithParam (155 ms) +[ RUN ] BiasLayerTest/0.TestForwardEltwiseInPlace +[ OK ] BiasLayerTest/0.TestForwardEltwiseInPlace (0 ms) +[ RUN ] BiasLayerTest/0.TestForwardBroadcastMiddle +[ OK ] BiasLayerTest/0.TestForwardBroadcastMiddle (1 ms) +[ RUN ] BiasLayerTest/0.TestForwardBroadcastMiddleWithParam +[ OK ] BiasLayerTest/0.TestForwardBroadcastMiddleWithParam (0 ms) +[ RUN ] BiasLayerTest/0.TestForwardBiasAxis2 +[ OK ] BiasLayerTest/0.TestForwardBiasAxis2 (0 ms) +[ RUN ] BiasLayerTest/0.TestBackwardBroadcastMiddleInPlace +[ OK ] BiasLayerTest/0.TestBackwardBroadcastMiddleInPlace (0 ms) +[ RUN ] BiasLayerTest/0.TestForwardBroadcastEnd +[ OK ] BiasLayerTest/0.TestForwardBroadcastEnd (0 ms) +[ RUN ] BiasLayerTest/0.TestGradientBroadcastEnd +[ OK ] BiasLayerTest/0.TestGradientBroadcastEnd (192 ms) +[ RUN ] BiasLayerTest/0.TestForwardBias +[ OK ] BiasLayerTest/0.TestForwardBias (0 ms) +[ RUN ] BiasLayerTest/0.TestGradientBroadcastBegin +[ OK ] BiasLayerTest/0.TestGradientBroadcastBegin (138 ms) +[ RUN ] BiasLayerTest/0.TestGradientBroadcastMiddle +[ OK ] BiasLayerTest/0.TestGradientBroadcastMiddle (156 ms) +[ RUN ] BiasLayerTest/0.TestGradientBias +[ OK ] BiasLayerTest/0.TestGradientBias (130 ms) +[ RUN ] BiasLayerTest/0.TestForwardBroadcastMiddleInPlace +[ OK ] BiasLayerTest/0.TestForwardBroadcastMiddleInPlace (0 ms) +[ RUN ] BiasLayerTest/0.TestGradientEltwiseWithParam +[ OK ] BiasLayerTest/0.TestGradientEltwiseWithParam (291 ms) +[ RUN ] BiasLayerTest/0.TestGradientBiasAxis2 +[ OK ] BiasLayerTest/0.TestGradientBiasAxis2 (126 ms) +[ RUN ] BiasLayerTest/0.TestForwardEltwiseWithParam +[ OK ] BiasLayerTest/0.TestForwardEltwiseWithParam (0 ms) +[----------] 20 tests from BiasLayerTest/0 (1195 ms total) + +[----------] 5 tests from MemoryDataLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] MemoryDataLayerTest/0.TestSetBatchSize +[ OK ] MemoryDataLayerTest/0.TestSetBatchSize (2 ms) +[ RUN ] MemoryDataLayerTest/0.TestForward +[ OK ] MemoryDataLayerTest/0.TestForward (49 ms) +[ RUN ] MemoryDataLayerTest/0.TestSetup +[ OK ] MemoryDataLayerTest/0.TestSetup (0 ms) +[ RUN ] MemoryDataLayerTest/0.AddMatVectorDefaultTransform +[ OK ] MemoryDataLayerTest/0.AddMatVectorDefaultTransform (2 ms) +[ RUN ] MemoryDataLayerTest/0.AddDatumVectorDefaultTransform +[ OK ] MemoryDataLayerTest/0.AddDatumVectorDefaultTransform (1 ms) +[----------] 5 tests from MemoryDataLayerTest/0 (56 ms total) + +[----------] 3 tests from MaxPoolingDropoutTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] MaxPoolingDropoutTest/0.TestSetup +[ OK ] MaxPoolingDropoutTest/0.TestSetup (0 ms) +[ RUN ] MaxPoolingDropoutTest/0.TestForward +[ OK ] MaxPoolingDropoutTest/0.TestForward (0 ms) +[ RUN ] MaxPoolingDropoutTest/0.TestBackward +[ OK ] MaxPoolingDropoutTest/0.TestBackward (0 ms) +[----------] 3 tests from MaxPoolingDropoutTest/0 (0 ms total) + +[----------] 22 tests from ScaleLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] ScaleLayerTest/1.TestGradientBroadcastMiddle +[ OK ] ScaleLayerTest/1.TestGradientBroadcastMiddle (175 ms) +[ RUN ] ScaleLayerTest/1.TestBackwardBroadcastMiddleInPlace +[ OK ] ScaleLayerTest/1.TestBackwardBroadcastMiddleInPlace (0 ms) +[ RUN ] ScaleLayerTest/1.TestGradientBroadcastBegin +[ OK ] ScaleLayerTest/1.TestGradientBroadcastBegin (150 ms) +[ RUN ] ScaleLayerTest/1.TestForwardBroadcastMiddleWithParamAndBias +[ OK ] ScaleLayerTest/1.TestForwardBroadcastMiddleWithParamAndBias (0 ms) +[ RUN ] ScaleLayerTest/1.TestForwardEltwiseWithParam +[ OK ] ScaleLayerTest/1.TestForwardEltwiseWithParam (0 ms) +[ RUN ] ScaleLayerTest/1.TestBackwardEltwiseInPlace +[ OK ] ScaleLayerTest/1.TestBackwardEltwiseInPlace (1 ms) +[ RUN ] ScaleLayerTest/1.TestForwardScale +[ OK ] ScaleLayerTest/1.TestForwardScale (0 ms) +[ RUN ] ScaleLayerTest/1.TestForwardBroadcastBegin +[ OK ] ScaleLayerTest/1.TestForwardBroadcastBegin (0 ms) +[ RUN ] ScaleLayerTest/1.TestForwardEltwiseInPlace +[ OK ] ScaleLayerTest/1.TestForwardEltwiseInPlace (0 ms) +[ RUN ] ScaleLayerTest/1.TestForwardBroadcastEnd +[ OK ] ScaleLayerTest/1.TestForwardBroadcastEnd (0 ms) +[ RUN ] ScaleLayerTest/1.TestGradientScaleAxis2 +[ OK ] ScaleLayerTest/1.TestGradientScaleAxis2 (134 ms) +[ RUN ] ScaleLayerTest/1.TestGradientEltwise +[ OK ] ScaleLayerTest/1.TestGradientEltwise (10 ms) +[ RUN ] ScaleLayerTest/1.TestGradientEltwiseWithParam +[ OK ] ScaleLayerTest/1.TestGradientEltwiseWithParam (572 ms) +[ RUN ] ScaleLayerTest/1.TestGradientBroadcastMiddleWithParam +[ OK ] ScaleLayerTest/1.TestGradientBroadcastMiddleWithParam (177 ms) +[ RUN ] ScaleLayerTest/1.TestForwardBroadcastMiddle +[ OK ] ScaleLayerTest/1.TestForwardBroadcastMiddle (0 ms) +[ RUN ] ScaleLayerTest/1.TestGradientScale +[ OK ] ScaleLayerTest/1.TestGradientScale (133 ms) +[ RUN ] ScaleLayerTest/1.TestForwardBroadcastMiddleInPlace +[ OK ] ScaleLayerTest/1.TestForwardBroadcastMiddleInPlace (0 ms) +[ RUN ] ScaleLayerTest/1.TestForwardBroadcastMiddleWithParam +[ OK ] ScaleLayerTest/1.TestForwardBroadcastMiddleWithParam (0 ms) +[ RUN ] ScaleLayerTest/1.TestForwardEltwise +[ OK ] ScaleLayerTest/1.TestForwardEltwise (0 ms) +[ RUN ] ScaleLayerTest/1.TestGradientScaleAndBias +[ OK ] ScaleLayerTest/1.TestGradientScaleAndBias (172 ms) +[ RUN ] ScaleLayerTest/1.TestForwardScaleAxis2 +[ OK ] ScaleLayerTest/1.TestForwardScaleAxis2 (0 ms) +[ RUN ] ScaleLayerTest/1.TestGradientBroadcastEnd +[ OK ] ScaleLayerTest/1.TestGradientBroadcastEnd (330 ms) +[----------] 22 tests from ScaleLayerTest/1 (1854 ms total) + +[----------] 26 tests from NetTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] NetTest/0.TestGetLayerByName +[ OK ] NetTest/0.TestGetLayerByName (2 ms) +[ RUN ] NetTest/0.TestUnsharedWeightsDataNet +[ OK ] NetTest/0.TestUnsharedWeightsDataNet (2 ms) +[ RUN ] NetTest/0.TestReshape +[ OK ] NetTest/0.TestReshape (2 ms) +[ RUN ] NetTest/0.TestBottomNeedBackwardForce +[ OK ] NetTest/0.TestBottomNeedBackwardForce (1 ms) +[ RUN ] NetTest/0.TestAllInOneNetDeploy +[ OK ] NetTest/0.TestAllInOneNetDeploy (1 ms) +[ RUN ] NetTest/0.TestBottomNeedBackwardTricky +[ OK ] NetTest/0.TestBottomNeedBackwardTricky (2 ms) +[ RUN ] NetTest/0.TestFromTo +[ OK ] NetTest/0.TestFromTo (4 ms) +[ RUN ] NetTest/0.TestBackwardWithAccuracyLayer +[ OK ] NetTest/0.TestBackwardWithAccuracyLayer (3 ms) +[ RUN ] NetTest/0.TestSharedWeightsDataNet +[ OK ] NetTest/0.TestSharedWeightsDataNet (1 ms) +[ RUN ] NetTest/0.TestBottomNeedBackward +[ OK ] NetTest/0.TestBottomNeedBackward (2 ms) +[ RUN ] NetTest/0.TestAllInOneNetVal +[ OK ] NetTest/0.TestAllInOneNetVal (1 ms) +[ RUN ] NetTest/0.TestSharedWeightsResume +[ OK ] NetTest/0.TestSharedWeightsResume (2 ms) +[ RUN ] NetTest/0.TestLossWeightMidNet +[ OK ] NetTest/0.TestLossWeightMidNet (10 ms) +[ RUN ] NetTest/0.TestLossWeight +[ OK ] NetTest/0.TestLossWeight (9 ms) +[ RUN ] NetTest/0.TestSharedWeightsDiffNet +[ OK ] NetTest/0.TestSharedWeightsDiffNet (1 ms) +[ RUN ] NetTest/0.TestHasLayer +[ OK ] NetTest/0.TestHasLayer (2 ms) +[ RUN ] NetTest/0.TestSkipPropagateDown +[ OK ] NetTest/0.TestSkipPropagateDown (3 ms) +[ RUN ] NetTest/0.TestUnsharedWeightsDiffNet +[ OK ] NetTest/0.TestUnsharedWeightsDiffNet (1 ms) +[ RUN ] NetTest/0.TestBottomNeedBackwardEuclideanForce +[ OK ] NetTest/0.TestBottomNeedBackwardEuclideanForce (1 ms) +[ RUN ] NetTest/0.TestParamPropagateDown +[ OK ] NetTest/0.TestParamPropagateDown (5 ms) +[ RUN ] NetTest/0.TestGetBlob +[ OK ] NetTest/0.TestGetBlob (2 ms) +[ RUN ] NetTest/0.TestHasBlob +[ OK ] NetTest/0.TestHasBlob (1 ms) +[ RUN ] NetTest/0.TestSharedWeightsUpdate +[ OK ] NetTest/0.TestSharedWeightsUpdate (3 ms) +[ RUN ] NetTest/0.TestComboLossWeight +[ OK ] NetTest/0.TestComboLossWeight (7 ms) +[ RUN ] NetTest/0.TestAllInOneNetTrain +[ OK ] NetTest/0.TestAllInOneNetTrain (2 ms) +[ RUN ] NetTest/0.TestForcePropagateDown +[ OK ] NetTest/0.TestForcePropagateDown (1 ms) +[----------] 26 tests from NetTest/0 (72 ms total) + +[----------] 9 tests from LSTMLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] LSTMLayerTest/0.TestGradientNonZeroCont +[ OK ] LSTMLayerTest/0.TestGradientNonZeroCont (542 ms) +[ RUN ] LSTMLayerTest/0.TestForward +[ OK ] LSTMLayerTest/0.TestForward (10 ms) +[ RUN ] LSTMLayerTest/0.TestSetUp +[ OK ] LSTMLayerTest/0.TestSetUp (2 ms) +[ RUN ] LSTMLayerTest/0.TestLSTMUnitGradientNonZeroCont +[ OK ] LSTMLayerTest/0.TestLSTMUnitGradientNonZeroCont (75 ms) +[ RUN ] LSTMLayerTest/0.TestGradient +[ OK ] LSTMLayerTest/0.TestGradient (545 ms) +[ RUN ] LSTMLayerTest/0.TestLSTMUnitGradient +[ OK ] LSTMLayerTest/0.TestLSTMUnitGradient (75 ms) +[ RUN ] LSTMLayerTest/0.TestLSTMUnitSetUp +[ OK ] LSTMLayerTest/0.TestLSTMUnitSetUp (0 ms) +[ RUN ] LSTMLayerTest/0.TestGradientNonZeroContBufferSize2 +[ OK ] LSTMLayerTest/0.TestGradientNonZeroContBufferSize2 (1037 ms) +[ RUN ] LSTMLayerTest/0.TestGradientNonZeroContBufferSize2WithStaticInput +[ OK ] LSTMLayerTest/0.TestGradientNonZeroContBufferSize2WithStaticInput (6112 ms) +[----------] 9 tests from LSTMLayerTest/0 (8398 ms total) + +[----------] 11 tests from AdaDeltaSolverTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] AdaDeltaSolverTest/1.TestLeastSquaresUpdateWithMomentumMultiIter +[ OK ] AdaDeltaSolverTest/1.TestLeastSquaresUpdateWithMomentumMultiIter (115 ms) +[ RUN ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithWeightDecay +[ OK ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithWeightDecay (24 ms) +[ RUN ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithHalfMomentum +[ OK ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithHalfMomentum (46 ms) +[ RUN ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithEverythingShare +[ OK ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithEverythingShare (127 ms) +[ RUN ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdate +[ OK ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdate (24 ms) +[ RUN ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithMomentum +[ OK ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithMomentum (47 ms) +[ RUN ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithEverything +[ OK ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithEverything (118 ms) +[ RUN ] AdaDeltaSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum +[ OK ] AdaDeltaSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum (7 ms) +[ RUN ] AdaDeltaSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare +[ OK ] AdaDeltaSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare (9 ms) +[ RUN ] AdaDeltaSolverTest/1.TestSnapshot +[ OK ] AdaDeltaSolverTest/1.TestSnapshot (43 ms) +[ RUN ] AdaDeltaSolverTest/1.TestSnapshotShare +[ OK ] AdaDeltaSolverTest/1.TestSnapshotShare (55 ms) +[----------] 11 tests from AdaDeltaSolverTest/1 (615 ms total) + +[----------] 6 tests from XavierFillerTest/0, where TypeParam = float +[ RUN ] XavierFillerTest/0.TestFillAverage +[ OK ] XavierFillerTest/0.TestFillAverage (63 ms) +[ RUN ] XavierFillerTest/0.TestFill1D +[ OK ] XavierFillerTest/0.TestFill1D (0 ms) +[ RUN ] XavierFillerTest/0.TestFill2D +[ OK ] XavierFillerTest/0.TestFill2D (0 ms) +[ RUN ] XavierFillerTest/0.TestFill5D +[ OK ] XavierFillerTest/0.TestFill5D (0 ms) +[ RUN ] XavierFillerTest/0.TestFillFanOut +[ OK ] XavierFillerTest/0.TestFillFanOut (63 ms) +[ RUN ] XavierFillerTest/0.TestFillFanIn +[ OK ] XavierFillerTest/0.TestFillFanIn (63 ms) +[----------] 6 tests from XavierFillerTest/0 (189 ms total) + +[----------] 3 tests from DummyDataLayerTest/1, where TypeParam = double +[ RUN ] DummyDataLayerTest/1.TestOneTopConstant +[ OK ] DummyDataLayerTest/1.TestOneTopConstant (0 ms) +[ RUN ] DummyDataLayerTest/1.TestThreeTopConstantGaussianConstant +[ OK ] DummyDataLayerTest/1.TestThreeTopConstantGaussianConstant (0 ms) +[ RUN ] DummyDataLayerTest/1.TestTwoTopConstant +[ OK ] DummyDataLayerTest/1.TestTwoTopConstant (0 ms) +[----------] 3 tests from DummyDataLayerTest/1 (0 ms total) [----------] 3 tests from PaddingLayerUpgradeTest -[ RUN ] PaddingLayerUpgradeTest.TestTwoTops -[ OK ] PaddingLayerUpgradeTest.TestTwoTops (1 ms) -[ RUN ] PaddingLayerUpgradeTest.TestImageNet -[ OK ] PaddingLayerUpgradeTest.TestImageNet (4 ms) [ RUN ] PaddingLayerUpgradeTest.TestSimple [ OK ] PaddingLayerUpgradeTest.TestSimple (1 ms) +[ RUN ] PaddingLayerUpgradeTest.TestImageNet +[ OK ] PaddingLayerUpgradeTest.TestImageNet (4 ms) +[ RUN ] PaddingLayerUpgradeTest.TestTwoTops +[ OK ] PaddingLayerUpgradeTest.TestTwoTops (1 ms) [----------] 3 tests from PaddingLayerUpgradeTest (6 ms total) -[----------] 4 tests from BlobSimpleTest/0, where TypeParam = float -[ RUN ] BlobSimpleTest/0.TestInitialization -[ OK ] BlobSimpleTest/0.TestInitialization (0 ms) -[ RUN ] BlobSimpleTest/0.TestLegacyBlobProtoShapeEquals -[ OK ] BlobSimpleTest/0.TestLegacyBlobProtoShapeEquals (0 ms) -[ RUN ] BlobSimpleTest/0.TestReshape -[ OK ] BlobSimpleTest/0.TestReshape (0 ms) -[ RUN ] BlobSimpleTest/0.TestReshapeZero -[ OK ] BlobSimpleTest/0.TestReshapeZero (0 ms) -[----------] 4 tests from BlobSimpleTest/0 (0 ms total) +[----------] 8 tests from Im2colLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] Im2colLayerTest/0.TestDilatedGradientForceND +[ OK ] Im2colLayerTest/0.TestDilatedGradientForceND (2520 ms) +[ RUN ] Im2colLayerTest/0.TestSetup +[ OK ] Im2colLayerTest/0.TestSetup (0 ms) +[ RUN ] Im2colLayerTest/0.TestForward +[ OK ] Im2colLayerTest/0.TestForward (0 ms) +[ RUN ] Im2colLayerTest/0.TestGradient +[ OK ] Im2colLayerTest/0.TestGradient (428 ms) +[ RUN ] Im2colLayerTest/0.TestRectGradient +[ OK ] Im2colLayerTest/0.TestRectGradient (364 ms) +[ RUN ] Im2colLayerTest/0.TestGradientForceND +[ OK ] Im2colLayerTest/0.TestGradientForceND (833 ms) +[ RUN ] Im2colLayerTest/0.TestDilatedGradient +[ OK ] Im2colLayerTest/0.TestDilatedGradient (1285 ms) +[ RUN ] Im2colLayerTest/0.TestRect +[ OK ] Im2colLayerTest/0.TestRect (1 ms) +[----------] 8 tests from Im2colLayerTest/0 (5431 ms total) -[----------] 8 tests from SliceLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] SliceLayerTest/0.TestSetupChannels -[ OK ] SliceLayerTest/0.TestSetupChannels (0 ms) -[ RUN ] SliceLayerTest/0.TestGradientAcrossNum -[ OK ] SliceLayerTest/0.TestGradientAcrossNum (72 ms) -[ RUN ] SliceLayerTest/0.TestGradientAcrossChannels -[ OK ] SliceLayerTest/0.TestGradientAcrossChannels (82 ms) -[ RUN ] SliceLayerTest/0.TestSetupNum -[ OK ] SliceLayerTest/0.TestSetupNum (0 ms) -[ RUN ] SliceLayerTest/0.TestSliceAcrossNum -[ OK ] SliceLayerTest/0.TestSliceAcrossNum (1 ms) -[ RUN ] SliceLayerTest/0.TestTrivialSlice -[ OK ] SliceLayerTest/0.TestTrivialSlice (0 ms) -[ RUN ] SliceLayerTest/0.TestGradientTrivial -[ OK ] SliceLayerTest/0.TestGradientTrivial (18 ms) -[ RUN ] SliceLayerTest/0.TestSliceAcrossChannels -[ OK ] SliceLayerTest/0.TestSliceAcrossChannels (0 ms) -[----------] 8 tests from SliceLayerTest/0 (173 ms total) +[----------] 7 tests from CPUMathFunctionsTest/0, where TypeParam = float +[ RUN ] CPUMathFunctionsTest/0.TestNothing +[ OK ] CPUMathFunctionsTest/0.TestNothing (3 ms) +[ RUN ] CPUMathFunctionsTest/0.TestSign +[ OK ] CPUMathFunctionsTest/0.TestSign (6 ms) +[ RUN ] CPUMathFunctionsTest/0.TestSgnbit +[ OK ] CPUMathFunctionsTest/0.TestSgnbit (5 ms) +[ RUN ] CPUMathFunctionsTest/0.TestCopy +[ OK ] CPUMathFunctionsTest/0.TestCopy (4 ms) +[ RUN ] CPUMathFunctionsTest/0.TestFabs +[ OK ] CPUMathFunctionsTest/0.TestFabs (6 ms) +[ RUN ] CPUMathFunctionsTest/0.TestAsum +[ OK ] CPUMathFunctionsTest/0.TestAsum (4 ms) +[ RUN ] CPUMathFunctionsTest/0.TestScale +[ OK ] CPUMathFunctionsTest/0.TestScale (5 ms) +[----------] 7 tests from CPUMathFunctionsTest/0 (34 ms total) -[----------] 12 tests from NesterovSolverTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] NesterovSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum -[ OK ] NesterovSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum (6 ms) -[ RUN ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithWeightDecayMultiIter -[ OK ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithWeightDecayMultiIter (112 ms) -[ RUN ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdate -[ OK ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdate (23 ms) -[ RUN ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithEverything -[ OK ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithEverything (112 ms) -[ RUN ] NesterovSolverTest/0.TestSnapshot -[ OK ] NesterovSolverTest/0.TestSnapshot (34 ms) -[ RUN ] NesterovSolverTest/0.TestLeastSquaresUpdateWithMomentumMultiIter -[ OK ] NesterovSolverTest/0.TestLeastSquaresUpdateWithMomentumMultiIter (112 ms) -[ RUN ] NesterovSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare -[ OK ] NesterovSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare (7 ms) -[ RUN ] NesterovSolverTest/0.TestSnapshotShare -[ OK ] NesterovSolverTest/0.TestSnapshotShare (44 ms) -[ RUN ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithWeightDecay -[ OK ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithWeightDecay (22 ms) -[ RUN ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithEverythingShare -[ OK ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithEverythingShare (117 ms) -[ RUN ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithMomentum -[ OK ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateWithMomentum (45 ms) -[ RUN ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateLROneHundredth -[ OK ] NesterovSolverTest/0.TestNesterovLeastSquaresUpdateLROneHundredth (23 ms) -[----------] 12 tests from NesterovSolverTest/0 (658 ms total) +[----------] 2 tests from HDF5DataLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] HDF5DataLayerTest/0.TestSkip +[ OK ] HDF5DataLayerTest/0.TestSkip (42 ms) +[ RUN ] HDF5DataLayerTest/0.TestRead +[ OK ] HDF5DataLayerTest/0.TestRead (11 ms) +[----------] 2 tests from HDF5DataLayerTest/0 (55 ms total) + +[----------] 8 tests from AdamSolverTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] AdamSolverTest/1.TestAdamLeastSquaresUpdate +[ OK ] AdamSolverTest/1.TestAdamLeastSquaresUpdate (23 ms) +[ RUN ] AdamSolverTest/1.TestAdamLeastSquaresUpdateWithEverything +[ OK ] AdamSolverTest/1.TestAdamLeastSquaresUpdateWithEverything (117 ms) +[ RUN ] AdamSolverTest/1.TestAdamLeastSquaresUpdateWithEverythingShare +[ OK ] AdamSolverTest/1.TestAdamLeastSquaresUpdateWithEverythingShare (125 ms) +[ RUN ] AdamSolverTest/1.TestSnapshot +[ OK ] AdamSolverTest/1.TestSnapshot (37 ms) +[ RUN ] AdamSolverTest/1.TestSnapshotShare +[ OK ] AdamSolverTest/1.TestSnapshotShare (46 ms) +[ RUN ] AdamSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare +[ OK ] AdamSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare (8 ms) +[ RUN ] AdamSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum +[ OK ] AdamSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum (6 ms) +[ RUN ] AdamSolverTest/1.TestAdamLeastSquaresUpdateWithWeightDecay +[ OK ] AdamSolverTest/1.TestAdamLeastSquaresUpdateWithWeightDecay (22 ms) +[----------] 8 tests from AdamSolverTest/1 (384 ms total) + +[----------] 11 tests from AdaDeltaSolverTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithEverythingShare +[ OK ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithEverythingShare (114 ms) +[ RUN ] AdaDeltaSolverTest/0.TestLeastSquaresUpdateWithMomentumMultiIter +[ OK ] AdaDeltaSolverTest/0.TestLeastSquaresUpdateWithMomentumMultiIter (112 ms) +[ RUN ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithMomentum +[ OK ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithMomentum (43 ms) +[ RUN ] AdaDeltaSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum +[ OK ] AdaDeltaSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum (5 ms) +[ RUN ] AdaDeltaSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare +[ OK ] AdaDeltaSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare (8 ms) +[ RUN ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithEverything +[ OK ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithEverything (112 ms) +[ RUN ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdate +[ OK ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdate (22 ms) +[ RUN ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithWeightDecay +[ OK ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithWeightDecay (22 ms) +[ RUN ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithHalfMomentum +[ OK ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithHalfMomentum (45 ms) +[ RUN ] AdaDeltaSolverTest/0.TestSnapshotShare +[ OK ] AdaDeltaSolverTest/0.TestSnapshotShare (48 ms) +[ RUN ] AdaDeltaSolverTest/0.TestSnapshot +[ OK ] AdaDeltaSolverTest/0.TestSnapshot (38 ms) +[----------] 11 tests from AdaDeltaSolverTest/0 (570 ms total) [----------] 8 tests from RMSPropSolverTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] RMSPropSolverTest/0.TestRMSPropLeastSquaresUpdateWithWeightDecay -[ OK ] RMSPropSolverTest/0.TestRMSPropLeastSquaresUpdateWithWeightDecay (22 ms) +[ RUN ] RMSPropSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare +[ OK ] RMSPropSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare (7 ms) [ RUN ] RMSPropSolverTest/0.TestRMSPropLeastSquaresUpdateWithEverythingShare -[ OK ] RMSPropSolverTest/0.TestRMSPropLeastSquaresUpdateWithEverythingShare (123 ms) -[ RUN ] RMSPropSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum -[ OK ] RMSPropSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum (6 ms) +[ OK ] RMSPropSolverTest/0.TestRMSPropLeastSquaresUpdateWithEverythingShare (120 ms) [ RUN ] RMSPropSolverTest/0.TestSnapshot [ OK ] RMSPropSolverTest/0.TestSnapshot (38 ms) -[ RUN ] RMSPropSolverTest/0.TestRMSPropLeastSquaresUpdateWithEverything -[ OK ] RMSPropSolverTest/0.TestRMSPropLeastSquaresUpdateWithEverything (114 ms) [ RUN ] RMSPropSolverTest/0.TestSnapshotShare -[ OK ] RMSPropSolverTest/0.TestSnapshotShare (46 ms) +[ OK ] RMSPropSolverTest/0.TestSnapshotShare (47 ms) [ RUN ] RMSPropSolverTest/0.TestRMSPropLeastSquaresUpdateWithRmsDecay -[ OK ] RMSPropSolverTest/0.TestRMSPropLeastSquaresUpdateWithRmsDecay (115 ms) -[ RUN ] RMSPropSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare -[ OK ] RMSPropSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare (7 ms) -[----------] 8 tests from RMSPropSolverTest/0 (472 ms total) +[ OK ] RMSPropSolverTest/0.TestRMSPropLeastSquaresUpdateWithRmsDecay (113 ms) +[ RUN ] RMSPropSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum +[ OK ] RMSPropSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum (6 ms) +[ RUN ] RMSPropSolverTest/0.TestRMSPropLeastSquaresUpdateWithEverything +[ OK ] RMSPropSolverTest/0.TestRMSPropLeastSquaresUpdateWithEverything (113 ms) +[ RUN ] RMSPropSolverTest/0.TestRMSPropLeastSquaresUpdateWithWeightDecay +[ OK ] RMSPropSolverTest/0.TestRMSPropLeastSquaresUpdateWithWeightDecay (23 ms) +[----------] 8 tests from RMSPropSolverTest/0 (469 ms total) + +[----------] 8 tests from AdamSolverTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] AdamSolverTest/0.TestSnapshotShare +[ OK ] AdamSolverTest/0.TestSnapshotShare (49 ms) +[ RUN ] AdamSolverTest/0.TestAdamLeastSquaresUpdate +[ OK ] AdamSolverTest/0.TestAdamLeastSquaresUpdate (23 ms) +[ RUN ] AdamSolverTest/0.TestAdamLeastSquaresUpdateWithWeightDecay +[ OK ] AdamSolverTest/0.TestAdamLeastSquaresUpdateWithWeightDecay (22 ms) +[ RUN ] AdamSolverTest/0.TestAdamLeastSquaresUpdateWithEverything +[ OK ] AdamSolverTest/0.TestAdamLeastSquaresUpdateWithEverything (114 ms) +[ RUN ] AdamSolverTest/0.TestAdamLeastSquaresUpdateWithEverythingShare +[ OK ] AdamSolverTest/0.TestAdamLeastSquaresUpdateWithEverythingShare (120 ms) +[ RUN ] AdamSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum +[ OK ] AdamSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum (6 ms) +[ RUN ] AdamSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare +[ OK ] AdamSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare (8 ms) +[ RUN ] AdamSolverTest/0.TestSnapshot +[ OK ] AdamSolverTest/0.TestSnapshot (35 ms) +[----------] 8 tests from AdamSolverTest/0 (377 ms total) + +[----------] 5 tests from EmbedLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] EmbedLayerTest/1.TestSetUp +[ OK ] EmbedLayerTest/1.TestSetUp (0 ms) +[ RUN ] EmbedLayerTest/1.TestGradient +[ OK ] EmbedLayerTest/1.TestGradient (18 ms) +[ RUN ] EmbedLayerTest/1.TestForwardWithBias +[ OK ] EmbedLayerTest/1.TestForwardWithBias (0 ms) +[ RUN ] EmbedLayerTest/1.TestForward +[ OK ] EmbedLayerTest/1.TestForward (0 ms) +[ RUN ] EmbedLayerTest/1.TestGradientWithBias +[ OK ] EmbedLayerTest/1.TestGradientWithBias (24 ms) +[----------] 5 tests from EmbedLayerTest/1 (42 ms total) + +[----------] 7 tests from TileLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] TileLayerTest/1.TestTrivialGradient +[ OK ] TileLayerTest/1.TestTrivialGradient (122 ms) +[ RUN ] TileLayerTest/1.TestGradientChannels +[ OK ] TileLayerTest/1.TestGradientChannels (422 ms) +[ RUN ] TileLayerTest/1.TestForwardNum +[ OK ] TileLayerTest/1.TestForwardNum (1 ms) +[ RUN ] TileLayerTest/1.TestForwardChannels +[ OK ] TileLayerTest/1.TestForwardChannels (0 ms) +[ RUN ] TileLayerTest/1.TestTrivialSetup +[ OK ] TileLayerTest/1.TestTrivialSetup (0 ms) +[ RUN ] TileLayerTest/1.TestSetup +[ OK ] TileLayerTest/1.TestSetup (0 ms) +[ RUN ] TileLayerTest/1.TestGradientNum +[ OK ] TileLayerTest/1.TestGradientNum (403 ms) +[----------] 7 tests from TileLayerTest/1 (948 ms total) + +[----------] 9 tests from AdaGradSolverTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdateWithWeightDecay +[ OK ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdateWithWeightDecay (23 ms) +[ RUN ] AdaGradSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare +[ OK ] AdaGradSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare (7 ms) +[ RUN ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdate +[ OK ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdate (23 ms) +[ RUN ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdateWithEverythingShare +[ OK ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdateWithEverythingShare (121 ms) +[ RUN ] AdaGradSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum +[ OK ] AdaGradSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum (6 ms) +[ RUN ] AdaGradSolverTest/0.TestSnapshot +[ OK ] AdaGradSolverTest/0.TestSnapshot (36 ms) +[ RUN ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdateWithEverything +[ OK ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdateWithEverything (110 ms) +[ RUN ] AdaGradSolverTest/0.TestSnapshotShare +[ OK ] AdaGradSolverTest/0.TestSnapshotShare (46 ms) +[ RUN ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdateLROneHundredth +[ OK ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdateLROneHundredth (22 ms) +[----------] 9 tests from AdaGradSolverTest/0 (395 ms total) [----------] 10 tests from EltwiseLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] EltwiseLayerTest/1.TestStableProdGradient -[ OK ] EltwiseLayerTest/1.TestStableProdGradient (7 ms) -[ RUN ] EltwiseLayerTest/1.TestSumCoeff -[ OK ] EltwiseLayerTest/1.TestSumCoeff (0 ms) [ RUN ] EltwiseLayerTest/1.TestSetUp [ OK ] EltwiseLayerTest/1.TestSetUp (0 ms) -[ RUN ] EltwiseLayerTest/1.TestUnstableProdGradient -[ OK ] EltwiseLayerTest/1.TestUnstableProdGradient (8 ms) -[ RUN ] EltwiseLayerTest/1.TestProd -[ OK ] EltwiseLayerTest/1.TestProd (0 ms) -[ RUN ] EltwiseLayerTest/1.TestMaxGradient -[ OK ] EltwiseLayerTest/1.TestMaxGradient (8 ms) +[ RUN ] EltwiseLayerTest/1.TestStableProdGradient +[ OK ] EltwiseLayerTest/1.TestStableProdGradient (7 ms) [ RUN ] EltwiseLayerTest/1.TestMax [ OK ] EltwiseLayerTest/1.TestMax (0 ms) +[ RUN ] EltwiseLayerTest/1.TestProd +[ OK ] EltwiseLayerTest/1.TestProd (1 ms) [ RUN ] EltwiseLayerTest/1.TestSumGradient -[ OK ] EltwiseLayerTest/1.TestSumGradient (8 ms) -[ RUN ] EltwiseLayerTest/1.TestSumCoeffGradient -[ OK ] EltwiseLayerTest/1.TestSumCoeffGradient (7 ms) +[ OK ] EltwiseLayerTest/1.TestSumGradient (7 ms) +[ RUN ] EltwiseLayerTest/1.TestMaxGradient +[ OK ] EltwiseLayerTest/1.TestMaxGradient (8 ms) [ RUN ] EltwiseLayerTest/1.TestSum [ OK ] EltwiseLayerTest/1.TestSum (0 ms) +[ RUN ] EltwiseLayerTest/1.TestUnstableProdGradient +[ OK ] EltwiseLayerTest/1.TestUnstableProdGradient (8 ms) +[ RUN ] EltwiseLayerTest/1.TestSumCoeffGradient +[ OK ] EltwiseLayerTest/1.TestSumCoeffGradient (7 ms) +[ RUN ] EltwiseLayerTest/1.TestSumCoeff +[ OK ] EltwiseLayerTest/1.TestSumCoeff (1 ms) [----------] 10 tests from EltwiseLayerTest/1 (39 ms total) -[----------] 5 tests from ImageDataLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] ImageDataLayerTest/1.TestSpace -[ OK ] ImageDataLayerTest/1.TestSpace (80 ms) -[ RUN ] ImageDataLayerTest/1.TestRead -[ OK ] ImageDataLayerTest/1.TestRead (196 ms) -[ RUN ] ImageDataLayerTest/1.TestResize -[ OK ] ImageDataLayerTest/1.TestResize (192 ms) -[ RUN ] ImageDataLayerTest/1.TestReshape -[ OK ] ImageDataLayerTest/1.TestReshape (51 ms) -[ RUN ] ImageDataLayerTest/1.TestShuffle -[ OK ] ImageDataLayerTest/1.TestShuffle (196 ms) -[----------] 5 tests from ImageDataLayerTest/1 (717 ms total) +[----------] 6 tests from RNNLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] RNNLayerTest/0.TestGradientNonZeroContBufferSize2 +[ OK ] RNNLayerTest/0.TestGradientNonZeroContBufferSize2 (390 ms) +[ RUN ] RNNLayerTest/0.TestGradientNonZeroCont +[ OK ] RNNLayerTest/0.TestGradientNonZeroCont (193 ms) +[ RUN ] RNNLayerTest/0.TestSetUp +[ OK ] RNNLayerTest/0.TestSetUp (3 ms) +[ RUN ] RNNLayerTest/0.TestGradientNonZeroContBufferSize2WithStaticInput +[ OK ] RNNLayerTest/0.TestGradientNonZeroContBufferSize2WithStaticInput (1821 ms) +[ RUN ] RNNLayerTest/0.TestGradient +[ OK ] RNNLayerTest/0.TestGradient (189 ms) +[ RUN ] RNNLayerTest/0.TestForward +[ OK ] RNNLayerTest/0.TestForward (10 ms) +[----------] 6 tests from RNNLayerTest/0 (2606 ms total) + +[----------] 9 tests from InnerProductLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] InnerProductLayerTest/0.TestSetUpTransposeTrue +[ OK ] InnerProductLayerTest/0.TestSetUpTransposeTrue (1 ms) +[ RUN ] InnerProductLayerTest/0.TestForwardTranspose +[ OK ] InnerProductLayerTest/0.TestForwardTranspose (0 ms) +[ RUN ] InnerProductLayerTest/0.TestSetUp +[ OK ] InnerProductLayerTest/0.TestSetUp (0 ms) +[ RUN ] InnerProductLayerTest/0.TestForwardNoBatch +[ OK ] InnerProductLayerTest/0.TestForwardNoBatch (0 ms) +[ RUN ] InnerProductLayerTest/0.TestGradient +[ OK ] InnerProductLayerTest/0.TestGradient (205 ms) +[ RUN ] InnerProductLayerTest/0.TestForward +[ OK ] InnerProductLayerTest/0.TestForward (0 ms) +[ RUN ] InnerProductLayerTest/0.TestSetUpTransposeFalse +[ OK ] InnerProductLayerTest/0.TestSetUpTransposeFalse (0 ms) +[ RUN ] InnerProductLayerTest/0.TestGradientTranspose +[ OK ] InnerProductLayerTest/0.TestGradientTranspose (249 ms) +[ RUN ] InnerProductLayerTest/0.TestBackwardTranspose +[ OK ] InnerProductLayerTest/0.TestBackwardTranspose (0 ms) +[----------] 9 tests from InnerProductLayerTest/0 (456 ms total) + +[----------] 4 tests from BlobSimpleTest/1, where TypeParam = double +[ RUN ] BlobSimpleTest/1.TestReshapeZero +[ OK ] BlobSimpleTest/1.TestReshapeZero (0 ms) +[ RUN ] BlobSimpleTest/1.TestInitialization +[ OK ] BlobSimpleTest/1.TestInitialization (0 ms) +[ RUN ] BlobSimpleTest/1.TestReshape +[ OK ] BlobSimpleTest/1.TestReshape (0 ms) +[ RUN ] BlobSimpleTest/1.TestLegacyBlobProtoShapeEquals +[ OK ] BlobSimpleTest/1.TestLegacyBlobProtoShapeEquals (0 ms) +[----------] 4 tests from BlobSimpleTest/1 (1 ms total) + +[----------] 5 tests from DeconvolutionLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] DeconvolutionLayerTest/1.TestGradient +[ OK ] DeconvolutionLayerTest/1.TestGradient (1126 ms) +[ RUN ] DeconvolutionLayerTest/1.TestSimpleDeconvolution +[ OK ] DeconvolutionLayerTest/1.TestSimpleDeconvolution (1 ms) +[ RUN ] DeconvolutionLayerTest/1.TestSetup +[ OK ] DeconvolutionLayerTest/1.TestSetup (0 ms) +[ RUN ] DeconvolutionLayerTest/1.TestNDAgainst2D +[ OK ] DeconvolutionLayerTest/1.TestNDAgainst2D (1997 ms) +[ RUN ] DeconvolutionLayerTest/1.TestGradient3D +[ OK ] DeconvolutionLayerTest/1.TestGradient3D (418 ms) +[----------] 5 tests from DeconvolutionLayerTest/1 (3542 ms total) + +[----------] 2 tests from InfogainLossLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] InfogainLossLayerTest/1.TestInfogainLoss +[ OK ] InfogainLossLayerTest/1.TestInfogainLoss (0 ms) +[ RUN ] InfogainLossLayerTest/1.TestGradient +[ OK ] InfogainLossLayerTest/1.TestGradient (3 ms) +[----------] 2 tests from InfogainLossLayerTest/1 (3 ms total) + +[----------] 58 tests from NeuronLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] NeuronLayerTest/1.TestPReLUGradient +[ OK ] NeuronLayerTest/1.TestPReLUGradient (160 ms) +[ RUN ] NeuronLayerTest/1.TestClip +[ OK ] NeuronLayerTest/1.TestClip (0 ms) +[ RUN ] NeuronLayerTest/1.TestExpLayerBase2 +[ OK ] NeuronLayerTest/1.TestExpLayerBase2 (0 ms) +[ RUN ] NeuronLayerTest/1.TestReLUWithNegativeSlope +[ OK ] NeuronLayerTest/1.TestReLUWithNegativeSlope (0 ms) +[ RUN ] NeuronLayerTest/1.TestSigmoidGradient +[ OK ] NeuronLayerTest/1.TestSigmoidGradient (7 ms) +[ RUN ] NeuronLayerTest/1.TestSwishGradient +[ OK ] NeuronLayerTest/1.TestSwishGradient (7 ms) +[ RUN ] NeuronLayerTest/1.TestLogGradientBase2Scale3 +[ OK ] NeuronLayerTest/1.TestLogGradientBase2Scale3 (5 ms) +[ RUN ] NeuronLayerTest/1.TestELUasReLU +[ OK ] NeuronLayerTest/1.TestELUasReLU (1 ms) +[ RUN ] NeuronLayerTest/1.TestExpLayerWithShift +[ OK ] NeuronLayerTest/1.TestExpLayerWithShift (0 ms) +[ RUN ] NeuronLayerTest/1.TestDropoutGradientTest +[ OK ] NeuronLayerTest/1.TestDropoutGradientTest (2 ms) +[ RUN ] NeuronLayerTest/1.TestBNLL +[ OK ] NeuronLayerTest/1.TestBNLL (0 ms) +[ RUN ] NeuronLayerTest/1.TestSwishWithBeta +[ OK ] NeuronLayerTest/1.TestSwishWithBeta (1 ms) +[ RUN ] NeuronLayerTest/1.TestPReLUConsistencyReLU +[ OK ] NeuronLayerTest/1.TestPReLUConsistencyReLU (0 ms) +[ RUN ] NeuronLayerTest/1.TestSwishAsLinearGradient +[ OK ] NeuronLayerTest/1.TestSwishAsLinearGradient (4 ms) +[ RUN ] NeuronLayerTest/1.TestPReLUParam +[ OK ] NeuronLayerTest/1.TestPReLUParam (0 ms) +[ RUN ] NeuronLayerTest/1.TestReLU +[ OK ] NeuronLayerTest/1.TestReLU (0 ms) +[ RUN ] NeuronLayerTest/1.TestExpLayer +[ OK ] NeuronLayerTest/1.TestExpLayer (0 ms) +[ RUN ] NeuronLayerTest/1.TestExpGradientBase2 +[ OK ] NeuronLayerTest/1.TestExpGradientBase2 (5 ms) +[ RUN ] NeuronLayerTest/1.TestAbsGradient +[ OK ] NeuronLayerTest/1.TestAbsGradient (2 ms) +[ RUN ] NeuronLayerTest/1.TestExpGradient +[ OK ] NeuronLayerTest/1.TestExpGradient (5 ms) +[ RUN ] NeuronLayerTest/1.TestReLUGradientWithNegativeSlope +[ OK ] NeuronLayerTest/1.TestReLUGradientWithNegativeSlope (3 ms) +[ RUN ] NeuronLayerTest/1.TestLogLayerBase2Shift1 +[ OK ] NeuronLayerTest/1.TestLogLayerBase2Shift1 (0 ms) +[ RUN ] NeuronLayerTest/1.TestReLUGradient +[ OK ] NeuronLayerTest/1.TestReLUGradient (3 ms) +[ RUN ] NeuronLayerTest/1.TestLogLayer +[ OK ] NeuronLayerTest/1.TestLogLayer (0 ms) +[ RUN ] NeuronLayerTest/1.TestLogLayerBase2Scale3 +[ OK ] NeuronLayerTest/1.TestLogLayerBase2Scale3 (0 ms) +[ RUN ] NeuronLayerTest/1.TestExpGradientBase2Shift1 +[ OK ] NeuronLayerTest/1.TestExpGradientBase2Shift1 (5 ms) +[ RUN ] NeuronLayerTest/1.TestClipGradient +[ OK ] NeuronLayerTest/1.TestClipGradient (3 ms) +[ RUN ] NeuronLayerTest/1.TestExpLayerBase2Shift1Scale3 +[ OK ] NeuronLayerTest/1.TestExpLayerBase2Shift1Scale3 (0 ms) +[ RUN ] NeuronLayerTest/1.TestExpGradientBase2Scale3 +[ OK ] NeuronLayerTest/1.TestExpGradientBase2Scale3 (4 ms) +[ RUN ] NeuronLayerTest/1.TestLogGradientBase2Shift1 +[ OK ] NeuronLayerTest/1.TestLogGradientBase2Shift1 (6 ms) +[ RUN ] NeuronLayerTest/1.TestSwishWithBetaGradient +[ OK ] NeuronLayerTest/1.TestSwishWithBetaGradient (9 ms) +[ RUN ] NeuronLayerTest/1.TestLogGradientBase2 +[ OK ] NeuronLayerTest/1.TestLogGradientBase2 (6 ms) +[ RUN ] NeuronLayerTest/1.TestExpGradientBase2Shift1Scale3 +[ OK ] NeuronLayerTest/1.TestExpGradientBase2Shift1Scale3 (4 ms) +[ RUN ] NeuronLayerTest/1.TestELU +[ OK ] NeuronLayerTest/1.TestELU (0 ms) +[ RUN ] NeuronLayerTest/1.TestExpGradientWithShift +[ OK ] NeuronLayerTest/1.TestExpGradientWithShift (5 ms) +[ RUN ] NeuronLayerTest/1.TestLogLayerBase2Shift1Scale3 +[ OK ] NeuronLayerTest/1.TestLogLayerBase2Shift1Scale3 (0 ms) +[ RUN ] NeuronLayerTest/1.TestELUasReLUGradient +[ OK ] NeuronLayerTest/1.TestELUasReLUGradient (4 ms) +[ RUN ] NeuronLayerTest/1.TestPReLUGradientChannelShared +[ OK ] NeuronLayerTest/1.TestPReLUGradientChannelShared (159 ms) +[ RUN ] NeuronLayerTest/1.TestLogGradient +[ OK ] NeuronLayerTest/1.TestLogGradient (6 ms) +[ RUN ] NeuronLayerTest/1.TestDropoutThreeQuarters +[ OK ] NeuronLayerTest/1.TestDropoutThreeQuarters (0 ms) +[ RUN ] NeuronLayerTest/1.TestAbsVal +[ OK ] NeuronLayerTest/1.TestAbsVal (0 ms) +[ RUN ] NeuronLayerTest/1.TestPReLUForward +[ OK ] NeuronLayerTest/1.TestPReLUForward (0 ms) +[ RUN ] NeuronLayerTest/1.TestSwishAsLinear +[ OK ] NeuronLayerTest/1.TestSwishAsLinear (0 ms) +[ RUN ] NeuronLayerTest/1.TestDropoutGradient +[ OK ] NeuronLayerTest/1.TestDropoutGradient (4 ms) +[ RUN ] NeuronLayerTest/1.TestSwish +[ OK ] NeuronLayerTest/1.TestSwish (0 ms) +[ RUN ] NeuronLayerTest/1.TestSigmoid +[ OK ] NeuronLayerTest/1.TestSigmoid (0 ms) +[ RUN ] NeuronLayerTest/1.TestTanHGradient +[ OK ] NeuronLayerTest/1.TestTanHGradient (7 ms) +[ RUN ] NeuronLayerTest/1.TestPReLUForwardChannelShared +[ OK ] NeuronLayerTest/1.TestPReLUForwardChannelShared (0 ms) +[ RUN ] NeuronLayerTest/1.TestLogGradientBase2Shift1Scale3 +[ OK ] NeuronLayerTest/1.TestLogGradientBase2Shift1Scale3 (6 ms) +[ RUN ] NeuronLayerTest/1.TestDropoutTestPhase +[ OK ] NeuronLayerTest/1.TestDropoutTestPhase (0 ms) +[ RUN ] NeuronLayerTest/1.TestDropoutHalf +[ OK ] NeuronLayerTest/1.TestDropoutHalf (0 ms) +[ RUN ] NeuronLayerTest/1.TestExpLayerBase2Shift1 +[ OK ] NeuronLayerTest/1.TestExpLayerBase2Shift1 (0 ms) +[ RUN ] NeuronLayerTest/1.TestBNLLGradient +[ OK ] NeuronLayerTest/1.TestBNLLGradient (6 ms) +[ RUN ] NeuronLayerTest/1.TestLogLayerBase2 +[ OK ] NeuronLayerTest/1.TestLogLayerBase2 (0 ms) +[ RUN ] NeuronLayerTest/1.TestExpLayerBase2Scale3 +[ OK ] NeuronLayerTest/1.TestExpLayerBase2Scale3 (0 ms) +[ RUN ] NeuronLayerTest/1.TestTanH +[ OK ] NeuronLayerTest/1.TestTanH (0 ms) +[ RUN ] NeuronLayerTest/1.TestPReLUInPlace +[ OK ] NeuronLayerTest/1.TestPReLUInPlace (0 ms) +[ RUN ] NeuronLayerTest/1.TestELUGradient +[ OK ] NeuronLayerTest/1.TestELUGradient (4 ms) +[----------] 58 tests from NeuronLayerTest/1 (448 ms total) [----------] 12 tests from SGDSolverTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] SGDSolverTest/0.TestLeastSquaresUpdateWithEverything -[ OK ] SGDSolverTest/0.TestLeastSquaresUpdateWithEverything (126 ms) -[ RUN ] SGDSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare -[ OK ] SGDSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare (7 ms) +[ RUN ] SGDSolverTest/0.TestLeastSquaresUpdateWithWeightDecayMultiIter +[ OK ] SGDSolverTest/0.TestLeastSquaresUpdateWithWeightDecayMultiIter (114 ms) +[ RUN ] SGDSolverTest/0.TestLeastSquaresUpdateWithMomentumMultiIter +[ OK ] SGDSolverTest/0.TestLeastSquaresUpdateWithMomentumMultiIter (112 ms) [ RUN ] SGDSolverTest/0.TestLeastSquaresUpdateWithMomentum -[ OK ] SGDSolverTest/0.TestLeastSquaresUpdateWithMomentum (43 ms) +[ OK ] SGDSolverTest/0.TestLeastSquaresUpdateWithMomentum (45 ms) +[ RUN ] SGDSolverTest/0.TestLeastSquaresUpdateWithEverythingShare +[ OK ] SGDSolverTest/0.TestLeastSquaresUpdateWithEverythingShare (118 ms) [ RUN ] SGDSolverTest/0.TestSnapshot [ OK ] SGDSolverTest/0.TestSnapshot (34 ms) -[ RUN ] SGDSolverTest/0.TestLeastSquaresUpdate -[ OK ] SGDSolverTest/0.TestLeastSquaresUpdate (22 ms) +[ RUN ] SGDSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum +[ OK ] SGDSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum (6 ms) [ RUN ] SGDSolverTest/0.TestLeastSquaresUpdateLROneHundredth [ OK ] SGDSolverTest/0.TestLeastSquaresUpdateLROneHundredth (22 ms) +[ RUN ] SGDSolverTest/0.TestSnapshotShare +[ OK ] SGDSolverTest/0.TestSnapshotShare (46 ms) +[ RUN ] SGDSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare +[ OK ] SGDSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare (8 ms) +[ RUN ] SGDSolverTest/0.TestLeastSquaresUpdate +[ OK ] SGDSolverTest/0.TestLeastSquaresUpdate (23 ms) +[ RUN ] SGDSolverTest/0.TestLeastSquaresUpdateWithEverything +[ OK ] SGDSolverTest/0.TestLeastSquaresUpdateWithEverything (111 ms) [ RUN ] SGDSolverTest/0.TestLeastSquaresUpdateWithWeightDecay [ OK ] SGDSolverTest/0.TestLeastSquaresUpdateWithWeightDecay (44 ms) -[ RUN ] SGDSolverTest/0.TestSnapshotShare -[ OK ] SGDSolverTest/0.TestSnapshotShare (43 ms) -[ RUN ] SGDSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum -[ OK ] SGDSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum (5 ms) -[ RUN ] SGDSolverTest/0.TestLeastSquaresUpdateWithMomentumMultiIter -[ OK ] SGDSolverTest/0.TestLeastSquaresUpdateWithMomentumMultiIter (107 ms) -[ RUN ] SGDSolverTest/0.TestLeastSquaresUpdateWithWeightDecayMultiIter -[ OK ] SGDSolverTest/0.TestLeastSquaresUpdateWithWeightDecayMultiIter (108 ms) -[ RUN ] SGDSolverTest/0.TestLeastSquaresUpdateWithEverythingShare -[ OK ] SGDSolverTest/0.TestLeastSquaresUpdateWithEverythingShare (117 ms) -[----------] 12 tests from SGDSolverTest/0 (680 ms total) - -[----------] 6 tests from XavierFillerTest/0, where TypeParam = float -[ RUN ] XavierFillerTest/0.TestFillFanIn -[ OK ] XavierFillerTest/0.TestFillFanIn (63 ms) -[ RUN ] XavierFillerTest/0.TestFill2D -[ OK ] XavierFillerTest/0.TestFill2D (0 ms) -[ RUN ] XavierFillerTest/0.TestFillAverage -[ OK ] XavierFillerTest/0.TestFillAverage (63 ms) -[ RUN ] XavierFillerTest/0.TestFillFanOut -[ OK ] XavierFillerTest/0.TestFillFanOut (63 ms) -[ RUN ] XavierFillerTest/0.TestFill1D -[ OK ] XavierFillerTest/0.TestFill1D (0 ms) -[ RUN ] XavierFillerTest/0.TestFill5D -[ OK ] XavierFillerTest/0.TestFill5D (0 ms) -[----------] 6 tests from XavierFillerTest/0 (189 ms total) - -[----------] 2 tests from CommonTest -[ RUN ] CommonTest.TestRandSeedCPU -[ OK ] CommonTest.TestRandSeedCPU (0 ms) -[ RUN ] CommonTest.TestBrewMode -[ OK ] CommonTest.TestBrewMode (0 ms) -[----------] 2 tests from CommonTest (0 ms total) - -[----------] 8 tests from SplitLayerInsertionTest -[ RUN ] SplitLayerInsertionTest.TestNoInsertion2 -[ OK ] SplitLayerInsertionTest.TestNoInsertion2 (1 ms) -[ RUN ] SplitLayerInsertionTest.TestLossInsertion -[ OK ] SplitLayerInsertionTest.TestLossInsertion (1 ms) -[ RUN ] SplitLayerInsertionTest.TestInsertion -[ OK ] SplitLayerInsertionTest.TestInsertion (0 ms) -[ RUN ] SplitLayerInsertionTest.TestWithInPlace -[ OK ] SplitLayerInsertionTest.TestWithInPlace (1 ms) -[ RUN ] SplitLayerInsertionTest.TestInsertionTwoTop -[ OK ] SplitLayerInsertionTest.TestInsertionTwoTop (0 ms) -[ RUN ] SplitLayerInsertionTest.TestNoInsertion1 -[ OK ] SplitLayerInsertionTest.TestNoInsertion1 (1 ms) -[ RUN ] SplitLayerInsertionTest.TestNoInsertionWithInPlace -[ OK ] SplitLayerInsertionTest.TestNoInsertionWithInPlace (0 ms) -[ RUN ] SplitLayerInsertionTest.TestNoInsertionImageNet -[ OK ] SplitLayerInsertionTest.TestNoInsertionImageNet (4 ms) -[----------] 8 tests from SplitLayerInsertionTest (9 ms total) - -[----------] 2 tests from BatchReindexLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] BatchReindexLayerTest/1.TestForward -[ OK ] BatchReindexLayerTest/1.TestForward (0 ms) -[ RUN ] BatchReindexLayerTest/1.TestGradient -[ OK ] BatchReindexLayerTest/1.TestGradient (224 ms) -[----------] 2 tests from BatchReindexLayerTest/1 (224 ms total) - -[----------] 20 tests from FilterNetTest -[ RUN ] FilterNetTest.TestFilterInByMultipleStage2 -[ OK ] FilterNetTest.TestFilterInByMultipleStage2 (1 ms) -[ RUN ] FilterNetTest.TestFilterLeNetTrainTest -[ OK ] FilterNetTest.TestFilterLeNetTrainTest (13 ms) -[ RUN ] FilterNetTest.TestFilterOutByStage -[ OK ] FilterNetTest.TestFilterOutByStage (0 ms) -[ RUN ] FilterNetTest.TestFilterOutByMaxLevel -[ OK ] FilterNetTest.TestFilterOutByMaxLevel (0 ms) -[ RUN ] FilterNetTest.TestNoFilter -[ OK ] FilterNetTest.TestNoFilter (0 ms) -[ RUN ] FilterNetTest.TestFilterOutByMultipleStage -[ OK ] FilterNetTest.TestFilterOutByMultipleStage (1 ms) -[ RUN ] FilterNetTest.TestFilterOutByMinLevel -[ OK ] FilterNetTest.TestFilterOutByMinLevel (0 ms) -[ RUN ] FilterNetTest.TestFilterInByIncludeMultiRule -[ OK ] FilterNetTest.TestFilterInByIncludeMultiRule (1 ms) -[ RUN ] FilterNetTest.TestFilterOutByNotStage -[ OK ] FilterNetTest.TestFilterOutByNotStage (0 ms) -[ RUN ] FilterNetTest.TestFilterInByMinLevel -[ OK ] FilterNetTest.TestFilterInByMinLevel (0 ms) -[ RUN ] FilterNetTest.TestFilterInByStage -[ OK ] FilterNetTest.TestFilterInByStage (1 ms) -[ RUN ] FilterNetTest.TestFilterInByMinLevel2 -[ OK ] FilterNetTest.TestFilterInByMinLevel2 (0 ms) -[ RUN ] FilterNetTest.TestFilterInByMaxLevel2 -[ OK ] FilterNetTest.TestFilterInByMaxLevel2 (1 ms) -[ RUN ] FilterNetTest.TestFilterInByNotStage -[ OK ] FilterNetTest.TestFilterInByNotStage (0 ms) -[ RUN ] FilterNetTest.TestFilterInByMultipleStage -[ OK ] FilterNetTest.TestFilterInByMultipleStage (0 ms) -[ RUN ] FilterNetTest.TestFilterInOutByExcludeMultiRule -[ OK ] FilterNetTest.TestFilterInOutByExcludeMultiRule (1 ms) -[ RUN ] FilterNetTest.TestFilterInByStage2 -[ OK ] FilterNetTest.TestFilterInByStage2 (1 ms) -[ RUN ] FilterNetTest.TestFilterInOutByIncludeMultiRule -[ OK ] FilterNetTest.TestFilterInOutByIncludeMultiRule (1 ms) -[ RUN ] FilterNetTest.TestFilterInByMaxLevel -[ OK ] FilterNetTest.TestFilterInByMaxLevel (0 ms) -[ RUN ] FilterNetTest.TestFilterOutByStage2 -[ OK ] FilterNetTest.TestFilterOutByStage2 (12 ms) -[----------] 20 tests from FilterNetTest (33 ms total) - -[----------] 5 tests from EmbedLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] EmbedLayerTest/0.TestForward -[ OK ] EmbedLayerTest/0.TestForward (0 ms) -[ RUN ] EmbedLayerTest/0.TestGradientWithBias -[ OK ] EmbedLayerTest/0.TestGradientWithBias (26 ms) -[ RUN ] EmbedLayerTest/0.TestSetUp -[ OK ] EmbedLayerTest/0.TestSetUp (0 ms) -[ RUN ] EmbedLayerTest/0.TestForwardWithBias -[ OK ] EmbedLayerTest/0.TestForwardWithBias (0 ms) -[ RUN ] EmbedLayerTest/0.TestGradient -[ OK ] EmbedLayerTest/0.TestGradient (19 ms) -[----------] 5 tests from EmbedLayerTest/0 (45 ms total) - -[----------] 3 tests from FilterLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] FilterLayerTest/0.TestForward -[ OK ] FilterLayerTest/0.TestForward (0 ms) -[ RUN ] FilterLayerTest/0.TestReshape -[ OK ] FilterLayerTest/0.TestReshape (0 ms) -[ RUN ] FilterLayerTest/0.TestGradient -[ OK ] FilterLayerTest/0.TestGradient (493 ms) -[----------] 3 tests from FilterLayerTest/0 (494 ms total) - -[----------] 11 tests from AdaDeltaSolverTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithWeightDecay -[ OK ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithWeightDecay (23 ms) -[ RUN ] AdaDeltaSolverTest/1.TestSnapshotShare -[ OK ] AdaDeltaSolverTest/1.TestSnapshotShare (52 ms) -[ RUN ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithEverythingShare -[ OK ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithEverythingShare (125 ms) -[ RUN ] AdaDeltaSolverTest/1.TestSnapshot -[ OK ] AdaDeltaSolverTest/1.TestSnapshot (42 ms) -[ RUN ] AdaDeltaSolverTest/1.TestLeastSquaresUpdateWithMomentumMultiIter -[ OK ] AdaDeltaSolverTest/1.TestLeastSquaresUpdateWithMomentumMultiIter (117 ms) -[ RUN ] AdaDeltaSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum -[ OK ] AdaDeltaSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum (7 ms) -[ RUN ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithMomentum -[ OK ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithMomentum (45 ms) -[ RUN ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdate -[ OK ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdate (22 ms) -[ RUN ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithEverything -[ OK ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithEverything (116 ms) -[ RUN ] AdaDeltaSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare -[ OK ] AdaDeltaSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare (8 ms) -[ RUN ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithHalfMomentum -[ OK ] AdaDeltaSolverTest/1.TestAdaDeltaLeastSquaresUpdateWithHalfMomentum (45 ms) -[----------] 11 tests from AdaDeltaSolverTest/1 (602 ms total) - -[----------] 1 test from SolverTypeUpgradeTest -[ RUN ] SolverTypeUpgradeTest.TestSimple -[ OK ] SolverTypeUpgradeTest.TestSimple (1 ms) -[----------] 1 test from SolverTypeUpgradeTest (1 ms total) +[----------] 12 tests from SGDSolverTest/0 (683 ms total) -[----------] 10 tests from EltwiseLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] EltwiseLayerTest/0.TestSum -[ OK ] EltwiseLayerTest/0.TestSum (0 ms) -[ RUN ] EltwiseLayerTest/0.TestMax -[ OK ] EltwiseLayerTest/0.TestMax (0 ms) -[ RUN ] EltwiseLayerTest/0.TestSetUp -[ OK ] EltwiseLayerTest/0.TestSetUp (0 ms) -[ RUN ] EltwiseLayerTest/0.TestSumCoeffGradient -[ OK ] EltwiseLayerTest/0.TestSumCoeffGradient (7 ms) -[ RUN ] EltwiseLayerTest/0.TestUnstableProdGradient -[ OK ] EltwiseLayerTest/0.TestUnstableProdGradient (6 ms) -[ RUN ] EltwiseLayerTest/0.TestSumCoeff -[ OK ] EltwiseLayerTest/0.TestSumCoeff (1 ms) -[ RUN ] EltwiseLayerTest/0.TestSumGradient -[ OK ] EltwiseLayerTest/0.TestSumGradient (7 ms) -[ RUN ] EltwiseLayerTest/0.TestProd -[ OK ] EltwiseLayerTest/0.TestProd (0 ms) -[ RUN ] EltwiseLayerTest/0.TestStableProdGradient -[ OK ] EltwiseLayerTest/0.TestStableProdGradient (6 ms) -[ RUN ] EltwiseLayerTest/0.TestMaxGradient -[ OK ] EltwiseLayerTest/0.TestMaxGradient (8 ms) -[----------] 10 tests from EltwiseLayerTest/0 (36 ms total) - -[----------] 6 tests from MSRAFillerTest/1, where TypeParam = double -[ RUN ] MSRAFillerTest/1.TestFillFanOut -[ OK ] MSRAFillerTest/1.TestFillFanOut (136 ms) -[ RUN ] MSRAFillerTest/1.TestFill2D -[ OK ] MSRAFillerTest/1.TestFill2D (1 ms) -[ RUN ] MSRAFillerTest/1.TestFill5D -[ OK ] MSRAFillerTest/1.TestFill5D (0 ms) -[ RUN ] MSRAFillerTest/1.TestFillFanIn -[ OK ] MSRAFillerTest/1.TestFillFanIn (136 ms) -[ RUN ] MSRAFillerTest/1.TestFill1D -[ OK ] MSRAFillerTest/1.TestFill1D (0 ms) -[ RUN ] MSRAFillerTest/1.TestFillAverage -[ OK ] MSRAFillerTest/1.TestFillAverage (137 ms) -[----------] 6 tests from MSRAFillerTest/1 (410 ms total) - -[----------] 3 tests from FilterLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] FilterLayerTest/1.TestGradient -[ OK ] FilterLayerTest/1.TestGradient (437 ms) -[ RUN ] FilterLayerTest/1.TestReshape -[ OK ] FilterLayerTest/1.TestReshape (0 ms) -[ RUN ] FilterLayerTest/1.TestForward -[ OK ] FilterLayerTest/1.TestForward (0 ms) -[----------] 3 tests from FilterLayerTest/1 (437 ms total) - -[----------] 2 tests from EuclideanLossLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] EuclideanLossLayerTest/0.TestForward -[ OK ] EuclideanLossLayerTest/0.TestForward (0 ms) -[ RUN ] EuclideanLossLayerTest/0.TestGradient -[ OK ] EuclideanLossLayerTest/0.TestGradient (1 ms) -[----------] 2 tests from EuclideanLossLayerTest/0 (1 ms total) - -[----------] 10 tests from PowerLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] PowerLayerTest/0.TestPowerZeroGradient -[ OK ] PowerLayerTest/0.TestPowerZeroGradient (3 ms) -[ RUN ] PowerLayerTest/0.TestPowerTwoScaleHalfGradient -[ OK ] PowerLayerTest/0.TestPowerTwoScaleHalfGradient (5 ms) -[ RUN ] PowerLayerTest/0.TestPowerZero -[ OK ] PowerLayerTest/0.TestPowerZero (0 ms) -[ RUN ] PowerLayerTest/0.TestPowerOne -[ OK ] PowerLayerTest/0.TestPowerOne (0 ms) -[ RUN ] PowerLayerTest/0.TestPowerGradient -[ OK ] PowerLayerTest/0.TestPowerGradient (5 ms) -[ RUN ] PowerLayerTest/0.TestPowerTwoGradient -[ OK ] PowerLayerTest/0.TestPowerTwoGradient (5 ms) -[ RUN ] PowerLayerTest/0.TestPowerOneGradient -[ OK ] PowerLayerTest/0.TestPowerOneGradient (2 ms) -[ RUN ] PowerLayerTest/0.TestPowerGradientShiftZero -[ OK ] PowerLayerTest/0.TestPowerGradientShiftZero (5 ms) -[ RUN ] PowerLayerTest/0.TestPowerTwo -[ OK ] PowerLayerTest/0.TestPowerTwo (0 ms) -[ RUN ] PowerLayerTest/0.TestPower -[ OK ] PowerLayerTest/0.TestPower (0 ms) -[----------] 10 tests from PowerLayerTest/0 (25 ms total) - -[----------] 9 tests from LSTMLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] LSTMLayerTest/0.TestLSTMUnitSetUp -[ OK ] LSTMLayerTest/0.TestLSTMUnitSetUp (0 ms) -[ RUN ] LSTMLayerTest/0.TestGradientNonZeroContBufferSize2WithStaticInput -[ OK ] LSTMLayerTest/0.TestGradientNonZeroContBufferSize2WithStaticInput (5746 ms) -[ RUN ] LSTMLayerTest/0.TestForward -[ OK ] LSTMLayerTest/0.TestForward (9 ms) -[ RUN ] LSTMLayerTest/0.TestLSTMUnitGradientNonZeroCont -[ OK ] LSTMLayerTest/0.TestLSTMUnitGradientNonZeroCont (70 ms) -[ RUN ] LSTMLayerTest/0.TestGradient -[ OK ] LSTMLayerTest/0.TestGradient (466 ms) -[ RUN ] LSTMLayerTest/0.TestGradientNonZeroCont -[ OK ] LSTMLayerTest/0.TestGradientNonZeroCont (467 ms) -[ RUN ] LSTMLayerTest/0.TestGradientNonZeroContBufferSize2 -[ OK ] LSTMLayerTest/0.TestGradientNonZeroContBufferSize2 (883 ms) -[ RUN ] LSTMLayerTest/0.TestSetUp -[ OK ] LSTMLayerTest/0.TestSetUp (3 ms) -[ RUN ] LSTMLayerTest/0.TestLSTMUnitGradient -[ OK ] LSTMLayerTest/0.TestLSTMUnitGradient (70 ms) -[----------] 9 tests from LSTMLayerTest/0 (7715 ms total) - -[----------] 5 tests from SPPLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] SPPLayerTest/1.TestEqualOutputDims -[ OK ] SPPLayerTest/1.TestEqualOutputDims (0 ms) -[ RUN ] SPPLayerTest/1.TestForwardBackward -[ OK ] SPPLayerTest/1.TestForwardBackward (1 ms) -[ RUN ] SPPLayerTest/1.TestSetup -[ OK ] SPPLayerTest/1.TestSetup (0 ms) -[ RUN ] SPPLayerTest/1.TestGradient -[ OK ] SPPLayerTest/1.TestGradient (3904 ms) -[ RUN ] SPPLayerTest/1.TestEqualOutputDims2 -[ OK ] SPPLayerTest/1.TestEqualOutputDims2 (0 ms) -[----------] 5 tests from SPPLayerTest/1 (3905 ms total) - -[----------] 22 tests from ScaleLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] ScaleLayerTest/1.TestGradientEltwiseWithParam -[ OK ] ScaleLayerTest/1.TestGradientEltwiseWithParam (540 ms) -[ RUN ] ScaleLayerTest/1.TestForwardScaleAxis2 -[ OK ] ScaleLayerTest/1.TestForwardScaleAxis2 (0 ms) -[ RUN ] ScaleLayerTest/1.TestForwardBroadcastMiddleInPlace -[ OK ] ScaleLayerTest/1.TestForwardBroadcastMiddleInPlace (1 ms) -[ RUN ] ScaleLayerTest/1.TestBackwardBroadcastMiddleInPlace -[ OK ] ScaleLayerTest/1.TestBackwardBroadcastMiddleInPlace (0 ms) -[ RUN ] ScaleLayerTest/1.TestForwardBroadcastEnd -[ OK ] ScaleLayerTest/1.TestForwardBroadcastEnd (0 ms) -[ RUN ] ScaleLayerTest/1.TestGradientBroadcastEnd -[ OK ] ScaleLayerTest/1.TestGradientBroadcastEnd (319 ms) -[ RUN ] ScaleLayerTest/1.TestForwardBroadcastBegin -[ OK ] ScaleLayerTest/1.TestForwardBroadcastBegin (0 ms) -[ RUN ] ScaleLayerTest/1.TestForwardScale -[ OK ] ScaleLayerTest/1.TestForwardScale (0 ms) -[ RUN ] ScaleLayerTest/1.TestBackwardEltwiseInPlace -[ OK ] ScaleLayerTest/1.TestBackwardEltwiseInPlace (0 ms) -[ RUN ] ScaleLayerTest/1.TestGradientBroadcastBegin -[ OK ] ScaleLayerTest/1.TestGradientBroadcastBegin (137 ms) -[ RUN ] ScaleLayerTest/1.TestGradientEltwise -[ OK ] ScaleLayerTest/1.TestGradientEltwise (9 ms) -[ RUN ] ScaleLayerTest/1.TestGradientScaleAxis2 -[ OK ] ScaleLayerTest/1.TestGradientScaleAxis2 (124 ms) -[ RUN ] ScaleLayerTest/1.TestForwardBroadcastMiddleWithParam -[ OK ] ScaleLayerTest/1.TestForwardBroadcastMiddleWithParam (0 ms) -[ RUN ] ScaleLayerTest/1.TestForwardBroadcastMiddle -[ OK ] ScaleLayerTest/1.TestForwardBroadcastMiddle (1 ms) -[ RUN ] ScaleLayerTest/1.TestForwardEltwise -[ OK ] ScaleLayerTest/1.TestForwardEltwise (0 ms) -[ RUN ] ScaleLayerTest/1.TestGradientScaleAndBias -[ OK ] ScaleLayerTest/1.TestGradientScaleAndBias (157 ms) -[ RUN ] ScaleLayerTest/1.TestGradientBroadcastMiddle -[ OK ] ScaleLayerTest/1.TestGradientBroadcastMiddle (163 ms) -[ RUN ] ScaleLayerTest/1.TestGradientBroadcastMiddleWithParam -[ OK ] ScaleLayerTest/1.TestGradientBroadcastMiddleWithParam (165 ms) -[ RUN ] ScaleLayerTest/1.TestForwardEltwiseWithParam -[ OK ] ScaleLayerTest/1.TestForwardEltwiseWithParam (1 ms) -[ RUN ] ScaleLayerTest/1.TestForwardBroadcastMiddleWithParamAndBias -[ OK ] ScaleLayerTest/1.TestForwardBroadcastMiddleWithParamAndBias (0 ms) -[ RUN ] ScaleLayerTest/1.TestForwardEltwiseInPlace -[ OK ] ScaleLayerTest/1.TestForwardEltwiseInPlace (0 ms) -[ RUN ] ScaleLayerTest/1.TestGradientScale -[ OK ] ScaleLayerTest/1.TestGradientScale (123 ms) -[----------] 22 tests from ScaleLayerTest/1 (1741 ms total) - -[----------] 1 test from MultinomialLogisticLossLayerTest/1, where TypeParam = double -[ RUN ] MultinomialLogisticLossLayerTest/1.TestGradientCPU -[ OK ] MultinomialLogisticLossLayerTest/1.TestGradientCPU (1 ms) -[----------] 1 test from MultinomialLogisticLossLayerTest/1 (1 ms total) - -[----------] 11 tests from RandomNumberGeneratorTest/0, where TypeParam = float -[ RUN ] RandomNumberGeneratorTest/0.TestRngGaussian -[ OK ] RandomNumberGeneratorTest/0.TestRngGaussian (0 ms) -[ RUN ] RandomNumberGeneratorTest/0.TestRngBernoulli -[ OK ] RandomNumberGeneratorTest/0.TestRngBernoulli (0 ms) -[ RUN ] RandomNumberGeneratorTest/0.TestRngUniform -[ OK ] RandomNumberGeneratorTest/0.TestRngUniform (1 ms) -[ RUN ] RandomNumberGeneratorTest/0.TestRngUniformTimesUniform -[ OK ] RandomNumberGeneratorTest/0.TestRngUniformTimesUniform (0 ms) -[ RUN ] RandomNumberGeneratorTest/0.TestRngBernoulliTimesBernoulli -[ OK ] RandomNumberGeneratorTest/0.TestRngBernoulliTimesBernoulli (1 ms) -[ RUN ] RandomNumberGeneratorTest/0.TestRngUniform2 -[ OK ] RandomNumberGeneratorTest/0.TestRngUniform2 (0 ms) -[ RUN ] RandomNumberGeneratorTest/0.TestRngBernoulli2 -[ OK ] RandomNumberGeneratorTest/0.TestRngBernoulli2 (0 ms) -[ RUN ] RandomNumberGeneratorTest/0.TestRngUniformTimesBernoulli -[ OK ] RandomNumberGeneratorTest/0.TestRngUniformTimesBernoulli (1 ms) -[ RUN ] RandomNumberGeneratorTest/0.TestRngGaussianTimesGaussian -[ OK ] RandomNumberGeneratorTest/0.TestRngGaussianTimesGaussian (0 ms) -[ RUN ] RandomNumberGeneratorTest/0.TestRngGaussian2 -[ OK ] RandomNumberGeneratorTest/0.TestRngGaussian2 (1 ms) -[ RUN ] RandomNumberGeneratorTest/0.TestRngGaussianTimesBernoulli -[ OK ] RandomNumberGeneratorTest/0.TestRngGaussianTimesBernoulli (0 ms) -[----------] 11 tests from RandomNumberGeneratorTest/0 (4 ms total) - -[----------] 9 tests from AccuracyLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] AccuracyLayerTest/0.TestSetup -[ OK ] AccuracyLayerTest/0.TestSetup (0 ms) -[ RUN ] AccuracyLayerTest/0.TestSetupOutputPerClass -[ OK ] AccuracyLayerTest/0.TestSetupOutputPerClass (0 ms) -[ RUN ] AccuracyLayerTest/0.TestForwardWithSpatialAxes -[ OK ] AccuracyLayerTest/0.TestForwardWithSpatialAxes (1 ms) -[ RUN ] AccuracyLayerTest/0.TestForwardTopK -[ OK ] AccuracyLayerTest/0.TestForwardTopK (6 ms) -[ RUN ] AccuracyLayerTest/0.TestForwardPerClass -[ OK ] AccuracyLayerTest/0.TestForwardPerClass (1 ms) -[ RUN ] AccuracyLayerTest/0.TestForward -[ OK ] AccuracyLayerTest/0.TestForward (1 ms) -[ RUN ] AccuracyLayerTest/0.TestSetupTopK -[ OK ] AccuracyLayerTest/0.TestSetupTopK (0 ms) -[ RUN ] AccuracyLayerTest/0.TestForwardPerClassWithIgnoreLabel -[ OK ] AccuracyLayerTest/0.TestForwardPerClassWithIgnoreLabel (1 ms) -[ RUN ] AccuracyLayerTest/0.TestForwardIgnoreLabel -[ OK ] AccuracyLayerTest/0.TestForwardIgnoreLabel (1 ms) -[----------] 9 tests from AccuracyLayerTest/0 (11 ms total) +[----------] 1 test from HDF5OutputLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] HDF5OutputLayerTest/1.TestForward +[ OK ] HDF5OutputLayerTest/1.TestForward (5 ms) +[----------] 1 test from HDF5OutputLayerTest/1 (5 ms total) -[----------] 3 tests from SigmoidCrossEntropyLossLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] SigmoidCrossEntropyLossLayerTest/1.TestSigmoidCrossEntropyLoss -[ OK ] SigmoidCrossEntropyLossLayerTest/1.TestSigmoidCrossEntropyLoss (3 ms) -[ RUN ] SigmoidCrossEntropyLossLayerTest/1.TestGradient -[ OK ] SigmoidCrossEntropyLossLayerTest/1.TestGradient (1 ms) -[ RUN ] SigmoidCrossEntropyLossLayerTest/1.TestIgnoreGradient -[ OK ] SigmoidCrossEntropyLossLayerTest/1.TestIgnoreGradient (0 ms) -[----------] 3 tests from SigmoidCrossEntropyLossLayerTest/1 (5 ms total) +[----------] 3 tests from SyncedMemoryTest +[ RUN ] SyncedMemoryTest.TestInitialization +[ OK ] SyncedMemoryTest.TestInitialization (0 ms) +[ RUN ] SyncedMemoryTest.TestAllocationCPU +[ OK ] SyncedMemoryTest.TestAllocationCPU (0 ms) +[ RUN ] SyncedMemoryTest.TestCPUWrite +[ OK ] SyncedMemoryTest.TestCPUWrite (0 ms) +[----------] 3 tests from SyncedMemoryTest (0 ms total) -[----------] 2 tests from HingeLossLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] HingeLossLayerTest/0.TestGradientL1 -[ OK ] HingeLossLayerTest/0.TestGradientL1 (0 ms) -[ RUN ] HingeLossLayerTest/0.TestGradientL2 -[ OK ] HingeLossLayerTest/0.TestGradientL2 (1 ms) -[----------] 2 tests from HingeLossLayerTest/0 (1 ms total) +[----------] 5 tests from SPPLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] SPPLayerTest/0.TestEqualOutputDims2 +[ OK ] SPPLayerTest/0.TestEqualOutputDims2 (1 ms) +[ RUN ] SPPLayerTest/0.TestForwardBackward +[ OK ] SPPLayerTest/0.TestForwardBackward (0 ms) +[ RUN ] SPPLayerTest/0.TestSetup +[ OK ] SPPLayerTest/0.TestSetup (0 ms) +[ RUN ] SPPLayerTest/0.TestGradient +[ OK ] SPPLayerTest/0.TestGradient (3858 ms) +[ RUN ] SPPLayerTest/0.TestEqualOutputDims +[ OK ] SPPLayerTest/0.TestEqualOutputDims (0 ms) +[----------] 5 tests from SPPLayerTest/0 (3860 ms total) -[----------] 1 test from SolverFactoryTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] SolverFactoryTest/0.TestCreateSolver -[ OK ] SolverFactoryTest/0.TestCreateSolver (2 ms) -[----------] 1 test from SolverFactoryTest/0 (2 ms total) +[----------] 3 tests from BatchNormLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] BatchNormLayerTest/1.TestForwardInplace +[ OK ] BatchNormLayerTest/1.TestForwardInplace (0 ms) +[ RUN ] BatchNormLayerTest/1.TestForward +[ OK ] BatchNormLayerTest/1.TestForward (0 ms) +[ RUN ] BatchNormLayerTest/1.TestGradient +[ OK ] BatchNormLayerTest/1.TestGradient (319 ms) +[----------] 3 tests from BatchNormLayerTest/1 (319 ms total) -[----------] 9 tests from InnerProductLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] InnerProductLayerTest/1.TestSetUpTransposeFalse -[ OK ] InnerProductLayerTest/1.TestSetUpTransposeFalse (0 ms) -[ RUN ] InnerProductLayerTest/1.TestSetUpTransposeTrue -[ OK ] InnerProductLayerTest/1.TestSetUpTransposeTrue (0 ms) -[ RUN ] InnerProductLayerTest/1.TestForwardTranspose -[ OK ] InnerProductLayerTest/1.TestForwardTranspose (0 ms) -[ RUN ] InnerProductLayerTest/1.TestForwardNoBatch -[ OK ] InnerProductLayerTest/1.TestForwardNoBatch (0 ms) -[ RUN ] InnerProductLayerTest/1.TestForward -[ OK ] InnerProductLayerTest/1.TestForward (0 ms) -[ RUN ] InnerProductLayerTest/1.TestGradientTranspose -[ OK ] InnerProductLayerTest/1.TestGradientTranspose (243 ms) -[ RUN ] InnerProductLayerTest/1.TestGradient -[ OK ] InnerProductLayerTest/1.TestGradient (215 ms) -[ RUN ] InnerProductLayerTest/1.TestBackwardTranspose -[ OK ] InnerProductLayerTest/1.TestBackwardTranspose (0 ms) -[ RUN ] InnerProductLayerTest/1.TestSetUp -[ OK ] InnerProductLayerTest/1.TestSetUp (0 ms) -[----------] 9 tests from InnerProductLayerTest/1 (459 ms total) +[----------] 12 tests from ReshapeLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] ReshapeLayerTest/0.TestInsertSingletonAxesMiddle +[ OK ] ReshapeLayerTest/0.TestInsertSingletonAxesMiddle (0 ms) +[ RUN ] ReshapeLayerTest/0.TestFlattenMiddle +[ OK ] ReshapeLayerTest/0.TestFlattenMiddle (0 ms) +[ RUN ] ReshapeLayerTest/0.TestInferenceOfUnspecified +[ OK ] ReshapeLayerTest/0.TestInferenceOfUnspecified (0 ms) +[ RUN ] ReshapeLayerTest/0.TestInsertSingletonAxesEnd +[ OK ] ReshapeLayerTest/0.TestInsertSingletonAxesEnd (0 ms) +[ RUN ] ReshapeLayerTest/0.TestForwardAfterReshape +[ OK ] ReshapeLayerTest/0.TestForwardAfterReshape (0 ms) +[ RUN ] ReshapeLayerTest/0.TestFlattenValues +[ OK ] ReshapeLayerTest/0.TestFlattenValues (0 ms) +[ RUN ] ReshapeLayerTest/0.TestFlattenOutputSizes +[ OK ] ReshapeLayerTest/0.TestFlattenOutputSizes (0 ms) +[ RUN ] ReshapeLayerTest/0.TestInsertSingletonAxesStart +[ OK ] ReshapeLayerTest/0.TestInsertSingletonAxesStart (0 ms) +[ RUN ] ReshapeLayerTest/0.TestInferenceOfUnspecifiedWithStartAxis +[ OK ] ReshapeLayerTest/0.TestInferenceOfUnspecifiedWithStartAxis (0 ms) +[ RUN ] ReshapeLayerTest/0.TestGradient +[ OK ] ReshapeLayerTest/0.TestGradient (5 ms) +[ RUN ] ReshapeLayerTest/0.TestCopyDimensions +[ OK ] ReshapeLayerTest/0.TestCopyDimensions (0 ms) +[ RUN ] ReshapeLayerTest/0.TestForward +[ OK ] ReshapeLayerTest/0.TestForward (0 ms) +[----------] 12 tests from ReshapeLayerTest/0 (5 ms total) -[----------] 26 tests from NetTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] NetTest/1.TestBottomNeedBackwardForce -[ OK ] NetTest/1.TestBottomNeedBackwardForce (2 ms) -[ RUN ] NetTest/1.TestForcePropagateDown -[ OK ] NetTest/1.TestForcePropagateDown (1 ms) -[ RUN ] NetTest/1.TestReshape -[ OK ] NetTest/1.TestReshape (2 ms) -[ RUN ] NetTest/1.TestFromTo -[ OK ] NetTest/1.TestFromTo (4 ms) -[ RUN ] NetTest/1.TestParamPropagateDown -[ OK ] NetTest/1.TestParamPropagateDown (5 ms) -[ RUN ] NetTest/1.TestLossWeightMidNet -[ OK ] NetTest/1.TestLossWeightMidNet (9 ms) -[ RUN ] NetTest/1.TestGetLayerByName -[ OK ] NetTest/1.TestGetLayerByName (1 ms) -[ RUN ] NetTest/1.TestAllInOneNetVal -[ OK ] NetTest/1.TestAllInOneNetVal (2 ms) -[ RUN ] NetTest/1.TestSharedWeightsDataNet -[ OK ] NetTest/1.TestSharedWeightsDataNet (1 ms) -[ RUN ] NetTest/1.TestHasLayer -[ OK ] NetTest/1.TestHasLayer (1 ms) -[ RUN ] NetTest/1.TestBackwardWithAccuracyLayer -[ OK ] NetTest/1.TestBackwardWithAccuracyLayer (4 ms) -[ RUN ] NetTest/1.TestSharedWeightsDiffNet -[ OK ] NetTest/1.TestSharedWeightsDiffNet (1 ms) -[ RUN ] NetTest/1.TestUnsharedWeightsDiffNet -[ OK ] NetTest/1.TestUnsharedWeightsDiffNet (1 ms) -[ RUN ] NetTest/1.TestHasBlob -[ OK ] NetTest/1.TestHasBlob (2 ms) -[ RUN ] NetTest/1.TestSharedWeightsResume -[ OK ] NetTest/1.TestSharedWeightsResume (1 ms) -[ RUN ] NetTest/1.TestSkipPropagateDown -[ OK ] NetTest/1.TestSkipPropagateDown (2 ms) -[ RUN ] NetTest/1.TestBottomNeedBackwardEuclideanForce -[ OK ] NetTest/1.TestBottomNeedBackwardEuclideanForce (1 ms) -[ RUN ] NetTest/1.TestUnsharedWeightsDataNet -[ OK ] NetTest/1.TestUnsharedWeightsDataNet (1 ms) -[ RUN ] NetTest/1.TestGetBlob -[ OK ] NetTest/1.TestGetBlob (2 ms) -[ RUN ] NetTest/1.TestLossWeight -[ OK ] NetTest/1.TestLossWeight (8 ms) -[ RUN ] NetTest/1.TestBottomNeedBackwardTricky -[ OK ] NetTest/1.TestBottomNeedBackwardTricky (2 ms) -[ RUN ] NetTest/1.TestAllInOneNetTrain -[ OK ] NetTest/1.TestAllInOneNetTrain (1 ms) -[ RUN ] NetTest/1.TestAllInOneNetDeploy -[ OK ] NetTest/1.TestAllInOneNetDeploy (1 ms) -[ RUN ] NetTest/1.TestBottomNeedBackward -[ OK ] NetTest/1.TestBottomNeedBackward (2 ms) -[ RUN ] NetTest/1.TestComboLossWeight -[ OK ] NetTest/1.TestComboLossWeight (6 ms) -[ RUN ] NetTest/1.TestSharedWeightsUpdate -[ OK ] NetTest/1.TestSharedWeightsUpdate (1 ms) -[----------] 26 tests from NetTest/1 (66 ms total) +[----------] 3 tests from BlobMathTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] BlobMathTest/1.TestScaleData +[ OK ] BlobMathTest/1.TestScaleData (0 ms) +[ RUN ] BlobMathTest/1.TestSumOfSquares +[ OK ] BlobMathTest/1.TestSumOfSquares (0 ms) +[ RUN ] BlobMathTest/1.TestAsum +[ OK ] BlobMathTest/1.TestAsum (0 ms) +[----------] 3 tests from BlobMathTest/1 (0 ms total) -[----------] 4 tests from SoftmaxWithLossLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] SoftmaxWithLossLayerTest/1.TestGradient -[ OK ] SoftmaxWithLossLayerTest/1.TestGradient (18 ms) -[ RUN ] SoftmaxWithLossLayerTest/1.TestForwardIgnoreLabel -[ OK ] SoftmaxWithLossLayerTest/1.TestForwardIgnoreLabel (0 ms) -[ RUN ] SoftmaxWithLossLayerTest/1.TestGradientIgnoreLabel -[ OK ] SoftmaxWithLossLayerTest/1.TestGradientIgnoreLabel (18 ms) -[ RUN ] SoftmaxWithLossLayerTest/1.TestGradientUnnormalized -[ OK ] SoftmaxWithLossLayerTest/1.TestGradientUnnormalized (18 ms) -[----------] 4 tests from SoftmaxWithLossLayerTest/1 (54 ms total) +[----------] 3 tests from SigmoidCrossEntropyLossLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] SigmoidCrossEntropyLossLayerTest/0.TestGradient +[ OK ] SigmoidCrossEntropyLossLayerTest/0.TestGradient (1 ms) +[ RUN ] SigmoidCrossEntropyLossLayerTest/0.TestSigmoidCrossEntropyLoss +[ OK ] SigmoidCrossEntropyLossLayerTest/0.TestSigmoidCrossEntropyLoss (2 ms) +[ RUN ] SigmoidCrossEntropyLossLayerTest/0.TestIgnoreGradient +[ OK ] SigmoidCrossEntropyLossLayerTest/0.TestIgnoreGradient (0 ms) +[----------] 3 tests from SigmoidCrossEntropyLossLayerTest/0 (4 ms total) -[----------] 3 tests from TanHLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] TanHLayerTest/0.TestTanHGradient -[ OK ] TanHLayerTest/0.TestTanHGradient (7 ms) -[ RUN ] TanHLayerTest/0.TestTanH -[ OK ] TanHLayerTest/0.TestTanH (0 ms) -[ RUN ] TanHLayerTest/0.TestTanHOverflow -[ OK ] TanHLayerTest/0.TestTanHOverflow (1 ms) -[----------] 3 tests from TanHLayerTest/0 (8 ms total) +[----------] 3 tests from TanHLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] TanHLayerTest/1.TestTanHGradient +[ OK ] TanHLayerTest/1.TestTanHGradient (8 ms) +[ RUN ] TanHLayerTest/1.TestTanHOverflow +[ OK ] TanHLayerTest/1.TestTanHOverflow (0 ms) +[ RUN ] TanHLayerTest/1.TestTanH +[ OK ] TanHLayerTest/1.TestTanH (0 ms) +[----------] 3 tests from TanHLayerTest/1 (8 ms total) -[----------] 10 tests from ConcatLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] ConcatLayerTest/1.TestSetupChannels -[ OK ] ConcatLayerTest/1.TestSetupChannels (0 ms) -[ RUN ] ConcatLayerTest/1.TestSetupNum -[ OK ] ConcatLayerTest/1.TestSetupNum (0 ms) -[ RUN ] ConcatLayerTest/1.TestForwardChannels -[ OK ] ConcatLayerTest/1.TestForwardChannels (0 ms) -[ RUN ] ConcatLayerTest/1.TestGradientNum -[ OK ] ConcatLayerTest/1.TestGradientNum (8 ms) -[ RUN ] ConcatLayerTest/1.TestForwardTrivial -[ OK ] ConcatLayerTest/1.TestForwardTrivial (0 ms) -[ RUN ] ConcatLayerTest/1.TestGradientTrivial -[ OK ] ConcatLayerTest/1.TestGradientTrivial (4 ms) -[ RUN ] ConcatLayerTest/1.TestForwardNum -[ OK ] ConcatLayerTest/1.TestForwardNum (1 ms) -[ RUN ] ConcatLayerTest/1.TestGradientChannels -[ OK ] ConcatLayerTest/1.TestGradientChannels (6 ms) -[ RUN ] ConcatLayerTest/1.TestGradientChannelsBottomOneOnly -[ OK ] ConcatLayerTest/1.TestGradientChannelsBottomOneOnly (3 ms) -[ RUN ] ConcatLayerTest/1.TestSetupChannelsNegativeIndexing -[ OK ] ConcatLayerTest/1.TestSetupChannelsNegativeIndexing (0 ms) -[----------] 10 tests from ConcatLayerTest/1 (22 ms total) +[----------] 2 tests from InternalThreadTest +[ RUN ] InternalThreadTest.TestRandomSeed +[ OK ] InternalThreadTest.TestRandomSeed (1 ms) +[ RUN ] InternalThreadTest.TestStartAndExit +[ OK ] InternalThreadTest.TestStartAndExit (0 ms) +[----------] 2 tests from InternalThreadTest (1 ms total) [----------] 20 tests from BiasLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] BiasLayerTest/1.TestForwardEltwise -[ OK ] BiasLayerTest/1.TestForwardEltwise (1 ms) -[ RUN ] BiasLayerTest/1.TestForwardBias -[ OK ] BiasLayerTest/1.TestForwardBias (0 ms) -[ RUN ] BiasLayerTest/1.TestGradientEltwise -[ OK ] BiasLayerTest/1.TestGradientEltwise (5 ms) -[ RUN ] BiasLayerTest/1.TestForwardBroadcastMiddleWithParam -[ OK ] BiasLayerTest/1.TestForwardBroadcastMiddleWithParam (1 ms) -[ RUN ] BiasLayerTest/1.TestForwardBroadcastMiddleInPlace -[ OK ] BiasLayerTest/1.TestForwardBroadcastMiddleInPlace (0 ms) -[ RUN ] BiasLayerTest/1.TestForwardBroadcastMiddle -[ OK ] BiasLayerTest/1.TestForwardBroadcastMiddle (0 ms) [ RUN ] BiasLayerTest/1.TestForwardEltwiseInPlace [ OK ] BiasLayerTest/1.TestForwardEltwiseInPlace (0 ms) +[ RUN ] BiasLayerTest/1.TestGradientBroadcastMiddleWithParam +[ OK ] BiasLayerTest/1.TestGradientBroadcastMiddleWithParam (159 ms) +[ RUN ] BiasLayerTest/1.TestGradientBias +[ OK ] BiasLayerTest/1.TestGradientBias (135 ms) +[ RUN ] BiasLayerTest/1.TestForwardBiasAxis2 +[ OK ] BiasLayerTest/1.TestForwardBiasAxis2 (0 ms) +[ RUN ] BiasLayerTest/1.TestGradientEltwiseWithParam +[ OK ] BiasLayerTest/1.TestGradientEltwiseWithParam (312 ms) +[ RUN ] BiasLayerTest/1.TestBackwardBroadcastMiddleInPlace +[ OK ] BiasLayerTest/1.TestBackwardBroadcastMiddleInPlace (0 ms) +[ RUN ] BiasLayerTest/1.TestBackwardEltwiseInPlace +[ OK ] BiasLayerTest/1.TestBackwardEltwiseInPlace (0 ms) +[ RUN ] BiasLayerTest/1.TestGradientEltwise +[ OK ] BiasLayerTest/1.TestGradientEltwise (6 ms) [ RUN ] BiasLayerTest/1.TestForwardEltwiseWithParam [ OK ] BiasLayerTest/1.TestForwardEltwiseWithParam (0 ms) +[ RUN ] BiasLayerTest/1.TestForwardEltwise +[ OK ] BiasLayerTest/1.TestForwardEltwise (0 ms) +[ RUN ] BiasLayerTest/1.TestForwardBroadcastMiddle +[ OK ] BiasLayerTest/1.TestForwardBroadcastMiddle (0 ms) +[ RUN ] BiasLayerTest/1.TestGradientBiasAxis2 +[ OK ] BiasLayerTest/1.TestGradientBiasAxis2 (135 ms) +[ RUN ] BiasLayerTest/1.TestGradientBroadcastMiddle +[ OK ] BiasLayerTest/1.TestGradientBroadcastMiddle (157 ms) +[ RUN ] BiasLayerTest/1.TestForwardBroadcastMiddleWithParam +[ OK ] BiasLayerTest/1.TestForwardBroadcastMiddleWithParam (0 ms) +[ RUN ] BiasLayerTest/1.TestGradientBroadcastEnd +[ OK ] BiasLayerTest/1.TestGradientBroadcastEnd (191 ms) [ RUN ] BiasLayerTest/1.TestGradientBroadcastBegin -[ OK ] BiasLayerTest/1.TestGradientBroadcastBegin (137 ms) -[ RUN ] BiasLayerTest/1.TestGradientEltwiseWithParam -[ OK ] BiasLayerTest/1.TestGradientEltwiseWithParam (296 ms) +[ OK ] BiasLayerTest/1.TestGradientBroadcastBegin (146 ms) [ RUN ] BiasLayerTest/1.TestForwardBroadcastBegin [ OK ] BiasLayerTest/1.TestForwardBroadcastBegin (0 ms) -[ RUN ] BiasLayerTest/1.TestGradientBiasAxis2 -[ OK ] BiasLayerTest/1.TestGradientBiasAxis2 (130 ms) -[ RUN ] BiasLayerTest/1.TestBackwardEltwiseInPlace -[ OK ] BiasLayerTest/1.TestBackwardEltwiseInPlace (0 ms) -[ RUN ] BiasLayerTest/1.TestGradientBroadcastMiddleWithParam -[ OK ] BiasLayerTest/1.TestGradientBroadcastMiddleWithParam (153 ms) -[ RUN ] BiasLayerTest/1.TestBackwardBroadcastMiddleInPlace -[ OK ] BiasLayerTest/1.TestBackwardBroadcastMiddleInPlace (0 ms) -[ RUN ] BiasLayerTest/1.TestGradientBias -[ OK ] BiasLayerTest/1.TestGradientBias (129 ms) +[ RUN ] BiasLayerTest/1.TestForwardBroadcastMiddleInPlace +[ OK ] BiasLayerTest/1.TestForwardBroadcastMiddleInPlace (0 ms) [ RUN ] BiasLayerTest/1.TestForwardBroadcastEnd [ OK ] BiasLayerTest/1.TestForwardBroadcastEnd (1 ms) -[ RUN ] BiasLayerTest/1.TestGradientBroadcastEnd -[ OK ] BiasLayerTest/1.TestGradientBroadcastEnd (185 ms) -[ RUN ] BiasLayerTest/1.TestForwardBiasAxis2 -[ OK ] BiasLayerTest/1.TestForwardBiasAxis2 (0 ms) -[ RUN ] BiasLayerTest/1.TestGradientBroadcastMiddle -[ OK ] BiasLayerTest/1.TestGradientBroadcastMiddle (151 ms) -[----------] 20 tests from BiasLayerTest/1 (1190 ms total) +[ RUN ] BiasLayerTest/1.TestForwardBias +[ OK ] BiasLayerTest/1.TestForwardBias (0 ms) +[----------] 20 tests from BiasLayerTest/1 (1243 ms total) -[----------] 6 tests from FlattenLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] FlattenLayerTest/1.TestSetupWithAxis -[ OK ] FlattenLayerTest/1.TestSetupWithAxis (0 ms) -[ RUN ] FlattenLayerTest/1.TestGradient -[ OK ] FlattenLayerTest/1.TestGradient (4 ms) -[ RUN ] FlattenLayerTest/1.TestSetupWithStartAndEndAxis -[ OK ] FlattenLayerTest/1.TestSetupWithStartAndEndAxis (0 ms) -[ RUN ] FlattenLayerTest/1.TestSetup -[ OK ] FlattenLayerTest/1.TestSetup (0 ms) -[ RUN ] FlattenLayerTest/1.TestForward -[ OK ] FlattenLayerTest/1.TestForward (1 ms) -[ RUN ] FlattenLayerTest/1.TestSetupWithEndAxis -[ OK ] FlattenLayerTest/1.TestSetupWithEndAxis (0 ms) -[----------] 6 tests from FlattenLayerTest/1 (5 ms total) +[----------] 4 tests from ContrastiveLossLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] ContrastiveLossLayerTest/1.TestForward +[ OK ] ContrastiveLossLayerTest/1.TestForward (0 ms) +[ RUN ] ContrastiveLossLayerTest/1.TestGradient +[ OK ] ContrastiveLossLayerTest/1.TestGradient (403 ms) +[ RUN ] ContrastiveLossLayerTest/1.TestForwardLegacy +[ OK ] ContrastiveLossLayerTest/1.TestForwardLegacy (1 ms) +[ RUN ] ContrastiveLossLayerTest/1.TestGradientLegacy +[ OK ] ContrastiveLossLayerTest/1.TestGradientLegacy (374 ms) +[----------] 4 tests from ContrastiveLossLayerTest/1 (778 ms total) -[----------] 8 tests from LRNLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] LRNLayerTest/1.TestForwardAcrossChannels -[ OK ] LRNLayerTest/1.TestForwardAcrossChannels (0 ms) -[ RUN ] LRNLayerTest/1.TestGradientAcrossChannels -[ OK ] LRNLayerTest/1.TestGradientAcrossChannels (1373 ms) -[ RUN ] LRNLayerTest/1.TestSetupWithinChannel -[ OK ] LRNLayerTest/1.TestSetupWithinChannel (0 ms) -[ RUN ] LRNLayerTest/1.TestGradientWithinChannel -[ OK ] LRNLayerTest/1.TestGradientWithinChannel (1265 ms) -[ RUN ] LRNLayerTest/1.TestForwardAcrossChannelsLargeRegion -[ OK ] LRNLayerTest/1.TestForwardAcrossChannelsLargeRegion (1 ms) -[ RUN ] LRNLayerTest/1.TestGradientAcrossChannelsLargeRegion -[ OK ] LRNLayerTest/1.TestGradientAcrossChannelsLargeRegion (1714 ms) -[ RUN ] LRNLayerTest/1.TestForwardWithinChannel -[ OK ] LRNLayerTest/1.TestForwardWithinChannel (1 ms) -[ RUN ] LRNLayerTest/1.TestSetupAcrossChannels -[ OK ] LRNLayerTest/1.TestSetupAcrossChannels (0 ms) -[----------] 8 tests from LRNLayerTest/1 (4354 ms total) +[----------] 2 tests from HingeLossLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] HingeLossLayerTest/0.TestGradientL2 +[ OK ] HingeLossLayerTest/0.TestGradientL2 (1 ms) +[ RUN ] HingeLossLayerTest/0.TestGradientL1 +[ OK ] HingeLossLayerTest/0.TestGradientL1 (0 ms) +[----------] 2 tests from HingeLossLayerTest/0 (1 ms total) -[----------] 5 tests from DBTest/1, where TypeParam = caffe::TypeLMDB -[ RUN ] DBTest/1.TestNext -[ OK ] DBTest/1.TestNext (16 ms) -[ RUN ] DBTest/1.TestKeyValue -[ OK ] DBTest/1.TestKeyValue (16 ms) -[ RUN ] DBTest/1.TestSeekToFirst -[ OK ] DBTest/1.TestSeekToFirst (14 ms) -[ RUN ] DBTest/1.TestWrite -[ OK ] DBTest/1.TestWrite (15 ms) -[ RUN ] DBTest/1.TestGetDB -[ OK ] DBTest/1.TestGetDB (14 ms) -[----------] 5 tests from DBTest/1 (76 ms total) +[----------] 2 tests from BatchReindexLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] BatchReindexLayerTest/0.TestForward +[ OK ] BatchReindexLayerTest/0.TestForward (1 ms) +[ RUN ] BatchReindexLayerTest/0.TestGradient +[ OK ] BatchReindexLayerTest/0.TestGradient (206 ms) +[----------] 2 tests from BatchReindexLayerTest/0 (207 ms total) -[----------] 7 tests from TileLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] TileLayerTest/0.TestForwardChannels -[ OK ] TileLayerTest/0.TestForwardChannels (0 ms) -[ RUN ] TileLayerTest/0.TestGradientNum -[ OK ] TileLayerTest/0.TestGradientNum (356 ms) -[ RUN ] TileLayerTest/0.TestSetup -[ OK ] TileLayerTest/0.TestSetup (1 ms) -[ RUN ] TileLayerTest/0.TestForwardNum -[ OK ] TileLayerTest/0.TestForwardNum (0 ms) -[ RUN ] TileLayerTest/0.TestGradientChannels -[ OK ] TileLayerTest/0.TestGradientChannels (393 ms) -[ RUN ] TileLayerTest/0.TestTrivialSetup -[ OK ] TileLayerTest/0.TestTrivialSetup (0 ms) -[ RUN ] TileLayerTest/0.TestTrivialGradient -[ OK ] TileLayerTest/0.TestTrivialGradient (117 ms) -[----------] 7 tests from TileLayerTest/0 (868 ms total) +[----------] 5 tests from ImageDataLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] ImageDataLayerTest/0.TestReshape +[ OK ] ImageDataLayerTest/0.TestReshape (50 ms) +[ RUN ] ImageDataLayerTest/0.TestShuffle +[ OK ] ImageDataLayerTest/0.TestShuffle (183 ms) +[ RUN ] ImageDataLayerTest/0.TestSpace +[ OK ] ImageDataLayerTest/0.TestSpace (52 ms) +[ RUN ] ImageDataLayerTest/0.TestResize +[ OK ] ImageDataLayerTest/0.TestResize (180 ms) +[ RUN ] ImageDataLayerTest/0.TestRead +[ OK ] ImageDataLayerTest/0.TestRead (184 ms) +[----------] 5 tests from ImageDataLayerTest/0 (649 ms total) -[----------] 1 test from HDF5OutputLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] HDF5OutputLayerTest/1.TestForward -[ OK ] HDF5OutputLayerTest/1.TestForward (5 ms) -[----------] 1 test from HDF5OutputLayerTest/1 (5 ms total) +[----------] 5 tests from ImageDataLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] ImageDataLayerTest/1.TestSpace +[ OK ] ImageDataLayerTest/1.TestSpace (66 ms) +[ RUN ] ImageDataLayerTest/1.TestReshape +[ OK ] ImageDataLayerTest/1.TestReshape (53 ms) +[ RUN ] ImageDataLayerTest/1.TestShuffle +[ OK ] ImageDataLayerTest/1.TestShuffle (232 ms) +[ RUN ] ImageDataLayerTest/1.TestResize +[ OK ] ImageDataLayerTest/1.TestResize (192 ms) +[ RUN ] ImageDataLayerTest/1.TestRead +[ OK ] ImageDataLayerTest/1.TestRead (191 ms) +[----------] 5 tests from ImageDataLayerTest/1 (734 ms total) -[----------] 14 tests from DataLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] DataLayerTest/0.TestReshapeLevelDB -[ OK ] DataLayerTest/0.TestReshapeLevelDB (3 ms) -[ RUN ] DataLayerTest/0.TestReadCropTrainSequenceSeededLevelDB -[ OK ] DataLayerTest/0.TestReadCropTrainSequenceSeededLevelDB (2 ms) -[ RUN ] DataLayerTest/0.TestReadLevelDB -[ OK ] DataLayerTest/0.TestReadLevelDB (6 ms) -[ RUN ] DataLayerTest/0.TestReadLMDB -[ OK ] DataLayerTest/0.TestReadLMDB (4 ms) -[ RUN ] DataLayerTest/0.TestReshapeLMDB -[ OK ] DataLayerTest/0.TestReshapeLMDB (1 ms) -[ RUN ] DataLayerTest/0.TestReadCropTrainSequenceUnseededLevelDB -[ OK ] DataLayerTest/0.TestReadCropTrainSequenceUnseededLevelDB (3 ms) -[ RUN ] DataLayerTest/0.TestReadCropTestLevelDB -[ OK ] DataLayerTest/0.TestReadCropTestLevelDB (1 ms) -[ RUN ] DataLayerTest/0.TestReadCropTrainLMDB -[ OK ] DataLayerTest/0.TestReadCropTrainLMDB (1 ms) -[ RUN ] DataLayerTest/0.TestSkipLMDB -[ OK ] DataLayerTest/0.TestSkipLMDB (6 ms) -[ RUN ] DataLayerTest/0.TestReadCropTrainLevelDB -[ OK ] DataLayerTest/0.TestReadCropTrainLevelDB (2 ms) -[ RUN ] DataLayerTest/0.TestReadCropTrainSequenceSeededLMDB -[ OK ] DataLayerTest/0.TestReadCropTrainSequenceSeededLMDB (1 ms) -[ RUN ] DataLayerTest/0.TestReadCropTestLMDB -[ OK ] DataLayerTest/0.TestReadCropTestLMDB (1 ms) -[ RUN ] DataLayerTest/0.TestReadCropTrainSequenceUnseededLMDB -[ OK ] DataLayerTest/0.TestReadCropTrainSequenceUnseededLMDB (1 ms) -[ RUN ] DataLayerTest/0.TestSkipLevelDB -[ OK ] DataLayerTest/0.TestSkipLevelDB (13 ms) -[----------] 14 tests from DataLayerTest/0 (45 ms total) +[----------] 11 tests from PoolingLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] PoolingLayerTest/1.TestSetupPadded +[ OK ] PoolingLayerTest/1.TestSetupPadded (0 ms) +[ RUN ] PoolingLayerTest/1.TestGradientAvePadded +[ OK ] PoolingLayerTest/1.TestGradientAvePadded (1554 ms) +[ RUN ] PoolingLayerTest/1.TestForwardMax +[ OK ] PoolingLayerTest/1.TestForwardMax (0 ms) +[ RUN ] PoolingLayerTest/1.TestGradientAve +[ OK ] PoolingLayerTest/1.TestGradientAve (380 ms) +[ RUN ] PoolingLayerTest/1.TestForwardMaxPadded +[ OK ] PoolingLayerTest/1.TestForwardMaxPadded (0 ms) +[ RUN ] PoolingLayerTest/1.TestGradientMaxTopMask +[ OK ] PoolingLayerTest/1.TestGradientMaxTopMask (980 ms) +[ RUN ] PoolingLayerTest/1.TestGradientMax +[ OK ] PoolingLayerTest/1.TestGradientMax (1232 ms) +[ RUN ] PoolingLayerTest/1.TestSetup +[ OK ] PoolingLayerTest/1.TestSetup (0 ms) +[ RUN ] PoolingLayerTest/1.TestForwardAve +[ OK ] PoolingLayerTest/1.TestForwardAve (0 ms) +[ RUN ] PoolingLayerTest/1.TestForwardMaxTopMask +[ OK ] PoolingLayerTest/1.TestForwardMaxTopMask (0 ms) +[ RUN ] PoolingLayerTest/1.TestSetupGlobalPooling +[ OK ] PoolingLayerTest/1.TestSetupGlobalPooling (0 ms) +[----------] 11 tests from PoolingLayerTest/1 (4146 ms total) -[----------] 1 test from ProtoTest -[ RUN ] ProtoTest.TestSerialization -Printing in binary format. +[----------] 12 tests from DataTransformTest/1, where TypeParam = double +[ RUN ] DataTransformTest/1.TestCropTest +[ OK ] DataTransformTest/1.TestCropTest (0 ms) +[ RUN ] DataTransformTest/1.TestMirrorTest +[ OK ] DataTransformTest/1.TestMirrorTest (0 ms) +[ RUN ] DataTransformTest/1.TestEmptyTransformUniquePixels +[ OK ] DataTransformTest/1.TestEmptyTransformUniquePixels (0 ms) +[ RUN ] DataTransformTest/1.TestMirrorTrain +[ OK ] DataTransformTest/1.TestMirrorTrain (0 ms) +[ RUN ] DataTransformTest/1.TestEmptyTransform +[ OK ] DataTransformTest/1.TestEmptyTransform (0 ms) +[ RUN ] DataTransformTest/1.TestCropMirrorTest +[ OK ] DataTransformTest/1.TestCropMirrorTest (0 ms) +[ RUN ] DataTransformTest/1.TestCropSize +[ OK ] DataTransformTest/1.TestCropSize (0 ms) +[ RUN ] DataTransformTest/1.TestCropTrain +[ OK ] DataTransformTest/1.TestCropTrain (1 ms) +[ RUN ] DataTransformTest/1.TestMeanValue +[ OK ] DataTransformTest/1.TestMeanValue (0 ms) +[ RUN ] DataTransformTest/1.TestMeanValues +[ OK ] DataTransformTest/1.TestMeanValues (0 ms) +[ RUN ] DataTransformTest/1.TestCropMirrorTrain +[ OK ] DataTransformTest/1.TestCropMirrorTrain (0 ms) +[ RUN ] DataTransformTest/1.TestMeanFile +[ OK ] DataTransformTest/1.TestMeanFile (0 ms) +[----------] 12 tests from DataTransformTest/1 (1 ms total) -testTest -Printing in text format. -name: "test" -type: "Test" +[----------] 11 tests from CropLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] CropLayerTest/0.TestSetupShapeAll +[ OK ] CropLayerTest/0.TestSetupShapeAll (0 ms) +[ RUN ] CropLayerTest/0.TestCropHW +[ OK ] CropLayerTest/0.TestCropHW (0 ms) +[ RUN ] CropLayerTest/0.TestCrop5DGradient +[ OK ] CropLayerTest/0.TestCrop5DGradient (2690 ms) +[ RUN ] CropLayerTest/0.TestCropAllGradient +[ OK ] CropLayerTest/0.TestCropAllGradient (438 ms) +[ RUN ] CropLayerTest/0.TestCropAll +[ OK ] CropLayerTest/0.TestCropAll (0 ms) +[ RUN ] CropLayerTest/0.TestSetupShapeDefault +[ OK ] CropLayerTest/0.TestSetupShapeDefault (0 ms) +[ RUN ] CropLayerTest/0.TestSetupShapeNegativeIndexing +[ OK ] CropLayerTest/0.TestSetupShapeNegativeIndexing (0 ms) +[ RUN ] CropLayerTest/0.TestCropAllOffset +[ OK ] CropLayerTest/0.TestCropAllOffset (0 ms) +[ RUN ] CropLayerTest/0.TestCropHWGradient +[ OK ] CropLayerTest/0.TestCropHWGradient (731 ms) +[ RUN ] CropLayerTest/0.TestCrop5D +[ OK ] CropLayerTest/0.TestCrop5D (1 ms) +[ RUN ] CropLayerTest/0.TestDimensionsCheck +[ OK ] CropLayerTest/0.TestDimensionsCheck (0 ms) +[----------] 11 tests from CropLayerTest/0 (3860 ms total) -[ OK ] ProtoTest.TestSerialization (0 ms) -[----------] 1 test from ProtoTest (0 ms total) +[----------] 3 tests from SplitLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] SplitLayerTest/1.Test +[ OK ] SplitLayerTest/1.Test (0 ms) +[ RUN ] SplitLayerTest/1.TestGradient +[ OK ] SplitLayerTest/1.TestGradient (9 ms) +[ RUN ] SplitLayerTest/1.TestSetup +[ OK ] SplitLayerTest/1.TestSetup (0 ms) +[----------] 3 tests from SplitLayerTest/1 (9 ms total) -[----------] 3 tests from SplitLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] SplitLayerTest/0.TestGradient -[ OK ] SplitLayerTest/0.TestGradient (8 ms) -[ RUN ] SplitLayerTest/0.Test -[ OK ] SplitLayerTest/0.Test (0 ms) -[ RUN ] SplitLayerTest/0.TestSetup -[ OK ] SplitLayerTest/0.TestSetup (0 ms) -[----------] 3 tests from SplitLayerTest/0 (8 ms total) +[----------] 3 tests from FilterLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] FilterLayerTest/0.TestGradient +[ OK ] FilterLayerTest/0.TestGradient (434 ms) +[ RUN ] FilterLayerTest/0.TestForward +[ OK ] FilterLayerTest/0.TestForward (1 ms) +[ RUN ] FilterLayerTest/0.TestReshape +[ OK ] FilterLayerTest/0.TestReshape (0 ms) +[----------] 3 tests from FilterLayerTest/0 (435 ms total) -[----------] 12 tests from ReshapeLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] ReshapeLayerTest/1.TestFlattenOutputSizes -[ OK ] ReshapeLayerTest/1.TestFlattenOutputSizes (0 ms) -[ RUN ] ReshapeLayerTest/1.TestInferenceOfUnspecified -[ OK ] ReshapeLayerTest/1.TestInferenceOfUnspecified (0 ms) -[ RUN ] ReshapeLayerTest/1.TestCopyDimensions -[ OK ] ReshapeLayerTest/1.TestCopyDimensions (0 ms) -[ RUN ] ReshapeLayerTest/1.TestInferenceOfUnspecifiedWithStartAxis -[ OK ] ReshapeLayerTest/1.TestInferenceOfUnspecifiedWithStartAxis (0 ms) -[ RUN ] ReshapeLayerTest/1.TestInsertSingletonAxesMiddle -[ OK ] ReshapeLayerTest/1.TestInsertSingletonAxesMiddle (0 ms) -[ RUN ] ReshapeLayerTest/1.TestForwardAfterReshape -[ OK ] ReshapeLayerTest/1.TestForwardAfterReshape (0 ms) -[ RUN ] ReshapeLayerTest/1.TestInsertSingletonAxesEnd -[ OK ] ReshapeLayerTest/1.TestInsertSingletonAxesEnd (0 ms) -[ RUN ] ReshapeLayerTest/1.TestFlattenMiddle -[ OK ] ReshapeLayerTest/1.TestFlattenMiddle (0 ms) -[ RUN ] ReshapeLayerTest/1.TestInsertSingletonAxesStart -[ OK ] ReshapeLayerTest/1.TestInsertSingletonAxesStart (0 ms) -[ RUN ] ReshapeLayerTest/1.TestGradient -[ OK ] ReshapeLayerTest/1.TestGradient (4 ms) -[ RUN ] ReshapeLayerTest/1.TestFlattenValues -[ OK ] ReshapeLayerTest/1.TestFlattenValues (0 ms) -[ RUN ] ReshapeLayerTest/1.TestForward -[ OK ] ReshapeLayerTest/1.TestForward (0 ms) -[----------] 12 tests from ReshapeLayerTest/1 (5 ms total) +[----------] 3 tests from ThresholdLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] ThresholdLayerTest/1.TestSetup +[ OK ] ThresholdLayerTest/1.TestSetup (0 ms) +[ RUN ] ThresholdLayerTest/1.Test2 +[ OK ] ThresholdLayerTest/1.Test2 (0 ms) +[ RUN ] ThresholdLayerTest/1.Test +[ OK ] ThresholdLayerTest/1.Test (0 ms) +[----------] 3 tests from ThresholdLayerTest/1 (0 ms total) -[----------] 1 test from LayerFactoryTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] LayerFactoryTest/1.TestCreateLayer -[ OK ] LayerFactoryTest/1.TestCreateLayer (2 ms) -[----------] 1 test from LayerFactoryTest/1 (2 ms total) +[----------] 4 tests from SoftmaxWithLossLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] SoftmaxWithLossLayerTest/1.TestForwardIgnoreLabel +[ OK ] SoftmaxWithLossLayerTest/1.TestForwardIgnoreLabel (0 ms) +[ RUN ] SoftmaxWithLossLayerTest/1.TestGradient +[ OK ] SoftmaxWithLossLayerTest/1.TestGradient (19 ms) +[ RUN ] SoftmaxWithLossLayerTest/1.TestGradientUnnormalized +[ OK ] SoftmaxWithLossLayerTest/1.TestGradientUnnormalized (19 ms) +[ RUN ] SoftmaxWithLossLayerTest/1.TestGradientIgnoreLabel +[ OK ] SoftmaxWithLossLayerTest/1.TestGradientIgnoreLabel (18 ms) +[----------] 4 tests from SoftmaxWithLossLayerTest/1 (56 ms total) -[----------] 11 tests from RandomNumberGeneratorTest/1, where TypeParam = double -[ RUN ] RandomNumberGeneratorTest/1.TestRngBernoulli -[ OK ] RandomNumberGeneratorTest/1.TestRngBernoulli (0 ms) -[ RUN ] RandomNumberGeneratorTest/1.TestRngGaussianTimesGaussian -[ OK ] RandomNumberGeneratorTest/1.TestRngGaussianTimesGaussian (1 ms) -[ RUN ] RandomNumberGeneratorTest/1.TestRngUniform -[ OK ] RandomNumberGeneratorTest/1.TestRngUniform (0 ms) -[ RUN ] RandomNumberGeneratorTest/1.TestRngUniformTimesBernoulli -[ OK ] RandomNumberGeneratorTest/1.TestRngUniformTimesBernoulli (1 ms) -[ RUN ] RandomNumberGeneratorTest/1.TestRngGaussianTimesBernoulli -[ OK ] RandomNumberGeneratorTest/1.TestRngGaussianTimesBernoulli (0 ms) -[ RUN ] RandomNumberGeneratorTest/1.TestRngBernoulli2 -[ OK ] RandomNumberGeneratorTest/1.TestRngBernoulli2 (1 ms) -[ RUN ] RandomNumberGeneratorTest/1.TestRngGaussian2 -[ OK ] RandomNumberGeneratorTest/1.TestRngGaussian2 (0 ms) -[ RUN ] RandomNumberGeneratorTest/1.TestRngUniformTimesUniform -[ OK ] RandomNumberGeneratorTest/1.TestRngUniformTimesUniform (0 ms) -[ RUN ] RandomNumberGeneratorTest/1.TestRngGaussian -[ OK ] RandomNumberGeneratorTest/1.TestRngGaussian (0 ms) -[ RUN ] RandomNumberGeneratorTest/1.TestRngBernoulliTimesBernoulli -[ OK ] RandomNumberGeneratorTest/1.TestRngBernoulliTimesBernoulli (0 ms) -[ RUN ] RandomNumberGeneratorTest/1.TestRngUniform2 -[ OK ] RandomNumberGeneratorTest/1.TestRngUniform2 (1 ms) -[----------] 11 tests from RandomNumberGeneratorTest/1 (5 ms total) +[----------] 2 tests from SoftmaxLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] SoftmaxLayerTest/0.TestForward +[ OK ] SoftmaxLayerTest/0.TestForward (0 ms) +[ RUN ] SoftmaxLayerTest/0.TestGradient +[ OK ] SoftmaxLayerTest/0.TestGradient (309 ms) +[----------] 2 tests from SoftmaxLayerTest/0 (309 ms total) [----------] 10 tests from PowerLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] PowerLayerTest/1.TestPowerOneGradient -[ OK ] PowerLayerTest/1.TestPowerOneGradient (2 ms) [ RUN ] PowerLayerTest/1.TestPowerGradient [ OK ] PowerLayerTest/1.TestPowerGradient (6 ms) +[ RUN ] PowerLayerTest/1.TestPowerZero +[ OK ] PowerLayerTest/1.TestPowerZero (0 ms) +[ RUN ] PowerLayerTest/1.TestPowerZeroGradient +[ OK ] PowerLayerTest/1.TestPowerZeroGradient (3 ms) [ RUN ] PowerLayerTest/1.TestPowerTwoGradient [ OK ] PowerLayerTest/1.TestPowerTwoGradient (6 ms) -[ RUN ] PowerLayerTest/1.TestPowerZeroGradient -[ OK ] PowerLayerTest/1.TestPowerZeroGradient (2 ms) -[ RUN ] PowerLayerTest/1.TestPowerOne -[ OK ] PowerLayerTest/1.TestPowerOne (0 ms) [ RUN ] PowerLayerTest/1.TestPowerTwoScaleHalfGradient [ OK ] PowerLayerTest/1.TestPowerTwoScaleHalfGradient (6 ms) [ RUN ] PowerLayerTest/1.TestPowerGradientShiftZero -[ OK ] PowerLayerTest/1.TestPowerGradientShiftZero (6 ms) -[ RUN ] PowerLayerTest/1.TestPowerZero -[ OK ] PowerLayerTest/1.TestPowerZero (0 ms) -[ RUN ] PowerLayerTest/1.TestPower -[ OK ] PowerLayerTest/1.TestPower (0 ms) +[ OK ] PowerLayerTest/1.TestPowerGradientShiftZero (7 ms) +[ RUN ] PowerLayerTest/1.TestPowerOne +[ OK ] PowerLayerTest/1.TestPowerOne (0 ms) +[ RUN ] PowerLayerTest/1.TestPowerOneGradient +[ OK ] PowerLayerTest/1.TestPowerOneGradient (2 ms) [ RUN ] PowerLayerTest/1.TestPowerTwo [ OK ] PowerLayerTest/1.TestPowerTwo (0 ms) -[----------] 10 tests from PowerLayerTest/1 (28 ms total) +[ RUN ] PowerLayerTest/1.TestPower +[ OK ] PowerLayerTest/1.TestPower (0 ms) +[----------] 10 tests from PowerLayerTest/1 (31 ms total) -[----------] 5 tests from EmbedLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] EmbedLayerTest/1.TestForward -[ OK ] EmbedLayerTest/1.TestForward (0 ms) -[ RUN ] EmbedLayerTest/1.TestGradient -[ OK ] EmbedLayerTest/1.TestGradient (18 ms) -[ RUN ] EmbedLayerTest/1.TestGradientWithBias -[ OK ] EmbedLayerTest/1.TestGradientWithBias (23 ms) -[ RUN ] EmbedLayerTest/1.TestForwardWithBias -[ OK ] EmbedLayerTest/1.TestForwardWithBias (0 ms) -[ RUN ] EmbedLayerTest/1.TestSetUp -[ OK ] EmbedLayerTest/1.TestSetUp (0 ms) -[----------] 5 tests from EmbedLayerTest/1 (41 ms total) +[----------] 3 tests from BlobMathTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] BlobMathTest/0.TestAsum +[ OK ] BlobMathTest/0.TestAsum (0 ms) +[ RUN ] BlobMathTest/0.TestSumOfSquares +[ OK ] BlobMathTest/0.TestSumOfSquares (0 ms) +[ RUN ] BlobMathTest/0.TestScaleData +[ OK ] BlobMathTest/0.TestScaleData (0 ms) +[----------] 3 tests from BlobMathTest/0 (0 ms total) -[----------] 1 test from LayerFactoryTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] LayerFactoryTest/0.TestCreateLayer -[ OK ] LayerFactoryTest/0.TestCreateLayer (3 ms) -[----------] 1 test from LayerFactoryTest/0 (3 ms total) +[----------] 11 tests from RandomNumberGeneratorTest/1, where TypeParam = double +[ RUN ] RandomNumberGeneratorTest/1.TestRngGaussianTimesGaussian +[ OK ] RandomNumberGeneratorTest/1.TestRngGaussianTimesGaussian (1 ms) +[ RUN ] RandomNumberGeneratorTest/1.TestRngBernoulli2 +[ OK ] RandomNumberGeneratorTest/1.TestRngBernoulli2 (0 ms) +[ RUN ] RandomNumberGeneratorTest/1.TestRngUniform2 +[ OK ] RandomNumberGeneratorTest/1.TestRngUniform2 (0 ms) +[ RUN ] RandomNumberGeneratorTest/1.TestRngUniformTimesUniform +[ OK ] RandomNumberGeneratorTest/1.TestRngUniformTimesUniform (1 ms) +[ RUN ] RandomNumberGeneratorTest/1.TestRngBernoulli +[ OK ] RandomNumberGeneratorTest/1.TestRngBernoulli (0 ms) +[ RUN ] RandomNumberGeneratorTest/1.TestRngGaussian +[ OK ] RandomNumberGeneratorTest/1.TestRngGaussian (0 ms) +[ RUN ] RandomNumberGeneratorTest/1.TestRngGaussian2 +[ OK ] RandomNumberGeneratorTest/1.TestRngGaussian2 (1 ms) +[ RUN ] RandomNumberGeneratorTest/1.TestRngUniformTimesBernoulli +[ OK ] RandomNumberGeneratorTest/1.TestRngUniformTimesBernoulli (0 ms) +[ RUN ] RandomNumberGeneratorTest/1.TestRngBernoulliTimesBernoulli +[ OK ] RandomNumberGeneratorTest/1.TestRngBernoulliTimesBernoulli (1 ms) +[ RUN ] RandomNumberGeneratorTest/1.TestRngUniform +[ OK ] RandomNumberGeneratorTest/1.TestRngUniform (0 ms) +[ RUN ] RandomNumberGeneratorTest/1.TestRngGaussianTimesBernoulli +[ OK ] RandomNumberGeneratorTest/1.TestRngGaussianTimesBernoulli (1 ms) +[----------] 11 tests from RandomNumberGeneratorTest/1 (5 ms total) -[----------] 3 tests from SplitLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] SplitLayerTest/1.TestGradient -[ OK ] SplitLayerTest/1.TestGradient (8 ms) -[ RUN ] SplitLayerTest/1.Test -[ OK ] SplitLayerTest/1.Test (0 ms) -[ RUN ] SplitLayerTest/1.TestSetup -[ OK ] SplitLayerTest/1.TestSetup (1 ms) -[----------] 3 tests from SplitLayerTest/1 (9 ms total) +[----------] 12 tests from NesterovSolverTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdate +[ OK ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdate (23 ms) +[ RUN ] NesterovSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum +[ OK ] NesterovSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum (6 ms) +[ RUN ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateLROneHundredth +[ OK ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateLROneHundredth (22 ms) +[ RUN ] NesterovSolverTest/1.TestLeastSquaresUpdateWithMomentumMultiIter +[ OK ] NesterovSolverTest/1.TestLeastSquaresUpdateWithMomentumMultiIter (114 ms) +[ RUN ] NesterovSolverTest/1.TestSnapshot +[ OK ] NesterovSolverTest/1.TestSnapshot (37 ms) +[ RUN ] NesterovSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare +[ OK ] NesterovSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare (8 ms) +[ RUN ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithWeightDecay +[ OK ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithWeightDecay (23 ms) +[ RUN ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithWeightDecayMultiIter +[ OK ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithWeightDecayMultiIter (112 ms) +[ RUN ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithMomentum +[ OK ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithMomentum (44 ms) +[ RUN ] NesterovSolverTest/1.TestSnapshotShare +[ OK ] NesterovSolverTest/1.TestSnapshotShare (45 ms) +[ RUN ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithEverything +[ OK ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithEverything (109 ms) +[ RUN ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithEverythingShare +[ OK ] NesterovSolverTest/1.TestNesterovLeastSquaresUpdateWithEverythingShare (116 ms) +[----------] 12 tests from NesterovSolverTest/1 (659 ms total) -[----------] 9 tests from LSTMLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] LSTMLayerTest/1.TestForward -[ OK ] LSTMLayerTest/1.TestForward (8 ms) -[ RUN ] LSTMLayerTest/1.TestGradientNonZeroCont -[ OK ] LSTMLayerTest/1.TestGradientNonZeroCont (462 ms) -[ RUN ] LSTMLayerTest/1.TestLSTMUnitGradientNonZeroCont -[ OK ] LSTMLayerTest/1.TestLSTMUnitGradientNonZeroCont (66 ms) -[ RUN ] LSTMLayerTest/1.TestGradientNonZeroContBufferSize2 -[ OK ] LSTMLayerTest/1.TestGradientNonZeroContBufferSize2 (873 ms) -[ RUN ] LSTMLayerTest/1.TestLSTMUnitGradient -[ OK ] LSTMLayerTest/1.TestLSTMUnitGradient (69 ms) -[ RUN ] LSTMLayerTest/1.TestGradientNonZeroContBufferSize2WithStaticInput -[ OK ] LSTMLayerTest/1.TestGradientNonZeroContBufferSize2WithStaticInput (5978 ms) -[ RUN ] LSTMLayerTest/1.TestGradient -[ OK ] LSTMLayerTest/1.TestGradient (485 ms) -[ RUN ] LSTMLayerTest/1.TestSetUp -[ OK ] LSTMLayerTest/1.TestSetUp (3 ms) -[ RUN ] LSTMLayerTest/1.TestLSTMUnitSetUp -[ OK ] LSTMLayerTest/1.TestLSTMUnitSetUp (0 ms) -[----------] 9 tests from LSTMLayerTest/1 (7944 ms total) +[----------] 8 tests from SplitLayerInsertionTest +[ RUN ] SplitLayerInsertionTest.TestNoInsertionImageNet +[ OK ] SplitLayerInsertionTest.TestNoInsertionImageNet (3 ms) +[ RUN ] SplitLayerInsertionTest.TestInsertion +[ OK ] SplitLayerInsertionTest.TestInsertion (0 ms) +[ RUN ] SplitLayerInsertionTest.TestNoInsertion1 +[ OK ] SplitLayerInsertionTest.TestNoInsertion1 (0 ms) +[ RUN ] SplitLayerInsertionTest.TestInsertionTwoTop +[ OK ] SplitLayerInsertionTest.TestInsertionTwoTop (1 ms) +[ RUN ] SplitLayerInsertionTest.TestWithInPlace +[ OK ] SplitLayerInsertionTest.TestWithInPlace (1 ms) +[ RUN ] SplitLayerInsertionTest.TestNoInsertionWithInPlace +[ OK ] SplitLayerInsertionTest.TestNoInsertionWithInPlace (0 ms) +[ RUN ] SplitLayerInsertionTest.TestLossInsertion +[ OK ] SplitLayerInsertionTest.TestLossInsertion (1 ms) +[ RUN ] SplitLayerInsertionTest.TestNoInsertion2 +[ OK ] SplitLayerInsertionTest.TestNoInsertion2 (0 ms) +[----------] 8 tests from SplitLayerInsertionTest (7 ms total) -[----------] 4 tests from ContrastiveLossLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] ContrastiveLossLayerTest/1.TestGradient -[ OK ] ContrastiveLossLayerTest/1.TestGradient (394 ms) -[ RUN ] ContrastiveLossLayerTest/1.TestForwardLegacy -[ OK ] ContrastiveLossLayerTest/1.TestForwardLegacy (0 ms) -[ RUN ] ContrastiveLossLayerTest/1.TestGradientLegacy -[ OK ] ContrastiveLossLayerTest/1.TestGradientLegacy (373 ms) -[ RUN ] ContrastiveLossLayerTest/1.TestForward -[ OK ] ContrastiveLossLayerTest/1.TestForward (1 ms) -[----------] 4 tests from ContrastiveLossLayerTest/1 (768 ms total) +[----------] 11 tests from CropLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] CropLayerTest/1.TestCrop5DGradient +[ OK ] CropLayerTest/1.TestCrop5DGradient (2620 ms) +[ RUN ] CropLayerTest/1.TestSetupShapeDefault +[ OK ] CropLayerTest/1.TestSetupShapeDefault (0 ms) +[ RUN ] CropLayerTest/1.TestCropAllGradient +[ OK ] CropLayerTest/1.TestCropAllGradient (432 ms) +[ RUN ] CropLayerTest/1.TestCropAll +[ OK ] CropLayerTest/1.TestCropAll (0 ms) +[ RUN ] CropLayerTest/1.TestCropHW +[ OK ] CropLayerTest/1.TestCropHW (0 ms) +[ RUN ] CropLayerTest/1.TestCropAllOffset +[ OK ] CropLayerTest/1.TestCropAllOffset (0 ms) +[ RUN ] CropLayerTest/1.TestCrop5D +[ OK ] CropLayerTest/1.TestCrop5D (0 ms) +[ RUN ] CropLayerTest/1.TestCropHWGradient +[ OK ] CropLayerTest/1.TestCropHWGradient (724 ms) +[ RUN ] CropLayerTest/1.TestSetupShapeNegativeIndexing +[ OK ] CropLayerTest/1.TestSetupShapeNegativeIndexing (0 ms) +[ RUN ] CropLayerTest/1.TestDimensionsCheck +[ OK ] CropLayerTest/1.TestDimensionsCheck (0 ms) +[ RUN ] CropLayerTest/1.TestSetupShapeAll +[ OK ] CropLayerTest/1.TestSetupShapeAll (0 ms) +[----------] 11 tests from CropLayerTest/1 (3777 ms total) -[----------] 8 tests from LRNLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] LRNLayerTest/0.TestSetupAcrossChannels -[ OK ] LRNLayerTest/0.TestSetupAcrossChannels (0 ms) -[ RUN ] LRNLayerTest/0.TestGradientAcrossChannels -[ OK ] LRNLayerTest/0.TestGradientAcrossChannels (1241 ms) -[ RUN ] LRNLayerTest/0.TestGradientAcrossChannelsLargeRegion -[ OK ] LRNLayerTest/0.TestGradientAcrossChannelsLargeRegion (1570 ms) -[ RUN ] LRNLayerTest/0.TestSetupWithinChannel -[ OK ] LRNLayerTest/0.TestSetupWithinChannel (0 ms) -[ RUN ] LRNLayerTest/0.TestForwardWithinChannel -[ OK ] LRNLayerTest/0.TestForwardWithinChannel (1 ms) -[ RUN ] LRNLayerTest/0.TestGradientWithinChannel -[ OK ] LRNLayerTest/0.TestGradientWithinChannel (1007 ms) -[ RUN ] LRNLayerTest/0.TestForwardAcrossChannelsLargeRegion -[ OK ] LRNLayerTest/0.TestForwardAcrossChannelsLargeRegion (0 ms) -[ RUN ] LRNLayerTest/0.TestForwardAcrossChannels -[ OK ] LRNLayerTest/0.TestForwardAcrossChannels (0 ms) -[----------] 8 tests from LRNLayerTest/0 (3819 ms total) +[----------] 9 tests from AccuracyLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] AccuracyLayerTest/1.TestForwardPerClassWithIgnoreLabel +[ OK ] AccuracyLayerTest/1.TestForwardPerClassWithIgnoreLabel (1 ms) +[ RUN ] AccuracyLayerTest/1.TestForward +[ OK ] AccuracyLayerTest/1.TestForward (1 ms) +[ RUN ] AccuracyLayerTest/1.TestSetupTopK +[ OK ] AccuracyLayerTest/1.TestSetupTopK (0 ms) +[ RUN ] AccuracyLayerTest/1.TestForwardPerClass +[ OK ] AccuracyLayerTest/1.TestForwardPerClass (1 ms) +[ RUN ] AccuracyLayerTest/1.TestForwardIgnoreLabel +[ OK ] AccuracyLayerTest/1.TestForwardIgnoreLabel (1 ms) +[ RUN ] AccuracyLayerTest/1.TestSetupOutputPerClass +[ OK ] AccuracyLayerTest/1.TestSetupOutputPerClass (0 ms) +[ RUN ] AccuracyLayerTest/1.TestForwardWithSpatialAxes +[ OK ] AccuracyLayerTest/1.TestForwardWithSpatialAxes (1 ms) +[ RUN ] AccuracyLayerTest/1.TestSetup +[ OK ] AccuracyLayerTest/1.TestSetup (0 ms) +[ RUN ] AccuracyLayerTest/1.TestForwardTopK +[ OK ] AccuracyLayerTest/1.TestForwardTopK (6 ms) +[----------] 9 tests from AccuracyLayerTest/1 (11 ms total) -[----------] 8 tests from Im2colLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] Im2colLayerTest/0.TestSetup -[ OK ] Im2colLayerTest/0.TestSetup (1 ms) -[ RUN ] Im2colLayerTest/0.TestDilatedGradientForceND -[ OK ] Im2colLayerTest/0.TestDilatedGradientForceND (2443 ms) -[ RUN ] Im2colLayerTest/0.TestRect -[ OK ] Im2colLayerTest/0.TestRect (0 ms) -[ RUN ] Im2colLayerTest/0.TestGradientForceND -[ OK ] Im2colLayerTest/0.TestGradientForceND (813 ms) -[ RUN ] Im2colLayerTest/0.TestRectGradient -[ OK ] Im2colLayerTest/0.TestRectGradient (355 ms) -[ RUN ] Im2colLayerTest/0.TestForward -[ OK ] Im2colLayerTest/0.TestForward (0 ms) -[ RUN ] Im2colLayerTest/0.TestGradient -[ OK ] Im2colLayerTest/0.TestGradient (421 ms) -[ RUN ] Im2colLayerTest/0.TestDilatedGradient -[ OK ] Im2colLayerTest/0.TestDilatedGradient (1239 ms) -[----------] 8 tests from Im2colLayerTest/0 (5273 ms total) +[----------] 8 tests from SliceLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] SliceLayerTest/0.TestSetupNum +[ OK ] SliceLayerTest/0.TestSetupNum (0 ms) +[ RUN ] SliceLayerTest/0.TestGradientAcrossNum +[ OK ] SliceLayerTest/0.TestGradientAcrossNum (59 ms) +[ RUN ] SliceLayerTest/0.TestSliceAcrossChannels +[ OK ] SliceLayerTest/0.TestSliceAcrossChannels (1 ms) +[ RUN ] SliceLayerTest/0.TestGradientAcrossChannels +[ OK ] SliceLayerTest/0.TestGradientAcrossChannels (67 ms) +[ RUN ] SliceLayerTest/0.TestTrivialSlice +[ OK ] SliceLayerTest/0.TestTrivialSlice (0 ms) +[ RUN ] SliceLayerTest/0.TestSetupChannels +[ OK ] SliceLayerTest/0.TestSetupChannels (0 ms) +[ RUN ] SliceLayerTest/0.TestGradientTrivial +[ OK ] SliceLayerTest/0.TestGradientTrivial (17 ms) +[ RUN ] SliceLayerTest/0.TestSliceAcrossNum +[ OK ] SliceLayerTest/0.TestSliceAcrossNum (0 ms) +[----------] 8 tests from SliceLayerTest/0 (145 ms total) -[----------] 2 tests from BilinearFillerTest/0, where TypeParam = float -[ RUN ] BilinearFillerTest/0.TestFillOdd -[ OK ] BilinearFillerTest/0.TestFillOdd (16 ms) -[ RUN ] BilinearFillerTest/0.TestFillEven -[ OK ] BilinearFillerTest/0.TestFillEven (12 ms) -[----------] 2 tests from BilinearFillerTest/0 (28 ms total) +[----------] 3 tests from DummyDataLayerTest/0, where TypeParam = float +[ RUN ] DummyDataLayerTest/0.TestOneTopConstant +[ OK ] DummyDataLayerTest/0.TestOneTopConstant (1 ms) +[ RUN ] DummyDataLayerTest/0.TestTwoTopConstant +[ OK ] DummyDataLayerTest/0.TestTwoTopConstant (0 ms) +[ RUN ] DummyDataLayerTest/0.TestThreeTopConstantGaussianConstant +[ OK ] DummyDataLayerTest/0.TestThreeTopConstantGaussianConstant (0 ms) +[----------] 3 tests from DummyDataLayerTest/0 (1 ms total) -[----------] 3 tests from SyncedMemoryTest -[ RUN ] SyncedMemoryTest.TestCPUWrite -[ OK ] SyncedMemoryTest.TestCPUWrite (0 ms) -[ RUN ] SyncedMemoryTest.TestInitialization -[ OK ] SyncedMemoryTest.TestInitialization (0 ms) -[ RUN ] SyncedMemoryTest.TestAllocationCPU -[ OK ] SyncedMemoryTest.TestAllocationCPU (0 ms) -[----------] 3 tests from SyncedMemoryTest (0 ms total) +[----------] 5 tests from DBTest/1, where TypeParam = caffe::TypeLMDB +[ RUN ] DBTest/1.TestGetDB +[ OK ] DBTest/1.TestGetDB (16 ms) +[ RUN ] DBTest/1.TestNext +[ OK ] DBTest/1.TestNext (16 ms) +[ RUN ] DBTest/1.TestWrite +[ OK ] DBTest/1.TestWrite (16 ms) +[ RUN ] DBTest/1.TestKeyValue +[ OK ] DBTest/1.TestKeyValue (16 ms) +[ RUN ] DBTest/1.TestSeekToFirst +[ OK ] DBTest/1.TestSeekToFirst (15 ms) +[----------] 5 tests from DBTest/1 (80 ms total) -[----------] 2 tests from SoftmaxLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] SoftmaxLayerTest/1.TestForward -[ OK ] SoftmaxLayerTest/1.TestForward (0 ms) -[ RUN ] SoftmaxLayerTest/1.TestGradient -[ OK ] SoftmaxLayerTest/1.TestGradient (335 ms) -[----------] 2 tests from SoftmaxLayerTest/1 (335 ms total) +[----------] 5 tests from MemoryDataLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] MemoryDataLayerTest/1.TestSetup +[ OK ] MemoryDataLayerTest/1.TestSetup (1 ms) +[ RUN ] MemoryDataLayerTest/1.AddDatumVectorDefaultTransform +[ OK ] MemoryDataLayerTest/1.AddDatumVectorDefaultTransform (1 ms) +[ RUN ] MemoryDataLayerTest/1.TestSetBatchSize +[ OK ] MemoryDataLayerTest/1.TestSetBatchSize (3 ms) +[ RUN ] MemoryDataLayerTest/1.AddMatVectorDefaultTransform +[ OK ] MemoryDataLayerTest/1.AddMatVectorDefaultTransform (2 ms) +[ RUN ] MemoryDataLayerTest/1.TestForward +[ OK ] MemoryDataLayerTest/1.TestForward (50 ms) +[----------] 5 tests from MemoryDataLayerTest/1 (58 ms total) -[----------] 26 tests from NetTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] NetTest/0.TestHasBlob -[ OK ] NetTest/0.TestHasBlob (1 ms) -[ RUN ] NetTest/0.TestBottomNeedBackward -[ OK ] NetTest/0.TestBottomNeedBackward (2 ms) -[ RUN ] NetTest/0.TestSharedWeightsUpdate -[ OK ] NetTest/0.TestSharedWeightsUpdate (2 ms) -[ RUN ] NetTest/0.TestSkipPropagateDown -[ OK ] NetTest/0.TestSkipPropagateDown (2 ms) -[ RUN ] NetTest/0.TestAllInOneNetVal -[ OK ] NetTest/0.TestAllInOneNetVal (2 ms) -[ RUN ] NetTest/0.TestComboLossWeight -[ OK ] NetTest/0.TestComboLossWeight (6 ms) -[ RUN ] NetTest/0.TestHasLayer -[ OK ] NetTest/0.TestHasLayer (2 ms) -[ RUN ] NetTest/0.TestLossWeightMidNet -[ OK ] NetTest/0.TestLossWeightMidNet (8 ms) -[ RUN ] NetTest/0.TestBottomNeedBackwardEuclideanForce -[ OK ] NetTest/0.TestBottomNeedBackwardEuclideanForce (1 ms) -[ RUN ] NetTest/0.TestSharedWeightsResume -[ OK ] NetTest/0.TestSharedWeightsResume (2 ms) -[ RUN ] NetTest/0.TestGetLayerByName -[ OK ] NetTest/0.TestGetLayerByName (2 ms) -[ RUN ] NetTest/0.TestUnsharedWeightsDiffNet -[ OK ] NetTest/0.TestUnsharedWeightsDiffNet (1 ms) -[ RUN ] NetTest/0.TestLossWeight -[ OK ] NetTest/0.TestLossWeight (8 ms) -[ RUN ] NetTest/0.TestGetBlob -[ OK ] NetTest/0.TestGetBlob (1 ms) -[ RUN ] NetTest/0.TestBottomNeedBackwardTricky -[ OK ] NetTest/0.TestBottomNeedBackwardTricky (2 ms) -[ RUN ] NetTest/0.TestParamPropagateDown -[ OK ] NetTest/0.TestParamPropagateDown (4 ms) -[ RUN ] NetTest/0.TestAllInOneNetDeploy -[ OK ] NetTest/0.TestAllInOneNetDeploy (1 ms) -[ RUN ] NetTest/0.TestForcePropagateDown -[ OK ] NetTest/0.TestForcePropagateDown (2 ms) -[ RUN ] NetTest/0.TestFromTo -[ OK ] NetTest/0.TestFromTo (4 ms) -[ RUN ] NetTest/0.TestReshape -[ OK ] NetTest/0.TestReshape (2 ms) -[ RUN ] NetTest/0.TestSharedWeightsDiffNet -[ OK ] NetTest/0.TestSharedWeightsDiffNet (1 ms) -[ RUN ] NetTest/0.TestBottomNeedBackwardForce -[ OK ] NetTest/0.TestBottomNeedBackwardForce (1 ms) -[ RUN ] NetTest/0.TestBackwardWithAccuracyLayer -[ OK ] NetTest/0.TestBackwardWithAccuracyLayer (3 ms) -[ RUN ] NetTest/0.TestAllInOneNetTrain -[ OK ] NetTest/0.TestAllInOneNetTrain (1 ms) -[ RUN ] NetTest/0.TestUnsharedWeightsDataNet -[ OK ] NetTest/0.TestUnsharedWeightsDataNet (2 ms) -[ RUN ] NetTest/0.TestSharedWeightsDataNet -[ OK ] NetTest/0.TestSharedWeightsDataNet (1 ms) -[----------] 26 tests from NetTest/0 (64 ms total) +[----------] 9 tests from InnerProductLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] InnerProductLayerTest/1.TestGradient +[ OK ] InnerProductLayerTest/1.TestGradient (211 ms) +[ RUN ] InnerProductLayerTest/1.TestForwardNoBatch +[ OK ] InnerProductLayerTest/1.TestForwardNoBatch (0 ms) +[ RUN ] InnerProductLayerTest/1.TestSetUpTransposeTrue +[ OK ] InnerProductLayerTest/1.TestSetUpTransposeTrue (0 ms) +[ RUN ] InnerProductLayerTest/1.TestBackwardTranspose +[ OK ] InnerProductLayerTest/1.TestBackwardTranspose (0 ms) +[ RUN ] InnerProductLayerTest/1.TestForwardTranspose +[ OK ] InnerProductLayerTest/1.TestForwardTranspose (0 ms) +[ RUN ] InnerProductLayerTest/1.TestSetUpTransposeFalse +[ OK ] InnerProductLayerTest/1.TestSetUpTransposeFalse (0 ms) +[ RUN ] InnerProductLayerTest/1.TestSetUp +[ OK ] InnerProductLayerTest/1.TestSetUp (0 ms) +[ RUN ] InnerProductLayerTest/1.TestGradientTranspose +[ OK ] InnerProductLayerTest/1.TestGradientTranspose (261 ms) +[ RUN ] InnerProductLayerTest/1.TestForward +[ OK ] InnerProductLayerTest/1.TestForward (0 ms) +[----------] 9 tests from InnerProductLayerTest/1 (473 ms total) -[----------] 15 tests from ConvolutionLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] ConvolutionLayerTest/0.TestGradient -[ OK ] ConvolutionLayerTest/0.TestGradient (125 ms) -[ RUN ] ConvolutionLayerTest/0.TestSimpleConvolutionGroup -[ OK ] ConvolutionLayerTest/0.TestSimpleConvolutionGroup (0 ms) -[ RUN ] ConvolutionLayerTest/0.TestNDAgainst2D -[ OK ] ConvolutionLayerTest/0.TestNDAgainst2D (664 ms) -[ RUN ] ConvolutionLayerTest/0.TestDilatedGradient -[ OK ] ConvolutionLayerTest/0.TestDilatedGradient (30 ms) -[ RUN ] ConvolutionLayerTest/0.TestSimple3DConvolution -[ OK ] ConvolutionLayerTest/0.TestSimple3DConvolution (2 ms) -[ RUN ] ConvolutionLayerTest/0.TestSetup -[ OK ] ConvolutionLayerTest/0.TestSetup (0 ms) -[ RUN ] ConvolutionLayerTest/0.TestDilated3DConvolution -[ OK ] ConvolutionLayerTest/0.TestDilated3DConvolution (15 ms) -[ RUN ] ConvolutionLayerTest/0.Test1x1Convolution -[ OK ] ConvolutionLayerTest/0.Test1x1Convolution (1 ms) -[ RUN ] ConvolutionLayerTest/0.Test0DConvolution -[ OK ] ConvolutionLayerTest/0.Test0DConvolution (0 ms) -[ RUN ] ConvolutionLayerTest/0.Test1x1Gradient -[ OK ] ConvolutionLayerTest/0.Test1x1Gradient (942 ms) -[ RUN ] ConvolutionLayerTest/0.TestGradient3D -[ OK ] ConvolutionLayerTest/0.TestGradient3D (1024 ms) -[ RUN ] ConvolutionLayerTest/0.TestDilatedConvolution -[ OK ] ConvolutionLayerTest/0.TestDilatedConvolution (3 ms) -[ RUN ] ConvolutionLayerTest/0.TestSimpleConvolution -[ OK ] ConvolutionLayerTest/0.TestSimpleConvolution (0 ms) -[ RUN ] ConvolutionLayerTest/0.TestGradientGroup -[ OK ] ConvolutionLayerTest/0.TestGradientGroup (34 ms) -[ RUN ] ConvolutionLayerTest/0.TestSobelConvolution -[ OK ] ConvolutionLayerTest/0.TestSobelConvolution (0 ms) -[----------] 15 tests from ConvolutionLayerTest/0 (2841 ms total) +[----------] 6 tests from MVNLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] MVNLayerTest/0.TestForwardMeanOnly +[ OK ] MVNLayerTest/0.TestForwardMeanOnly (1 ms) +[ RUN ] MVNLayerTest/0.TestForwardAcrossChannels +[ OK ] MVNLayerTest/0.TestForwardAcrossChannels (0 ms) +[ RUN ] MVNLayerTest/0.TestGradientAcrossChannels +[ OK ] MVNLayerTest/0.TestGradientAcrossChannels (430 ms) +[ RUN ] MVNLayerTest/0.TestForward +[ OK ] MVNLayerTest/0.TestForward (0 ms) +[ RUN ] MVNLayerTest/0.TestGradient +[ OK ] MVNLayerTest/0.TestGradient (427 ms) +[ RUN ] MVNLayerTest/0.TestGradientMeanOnly +[ OK ] MVNLayerTest/0.TestGradientMeanOnly (170 ms) +[----------] 6 tests from MVNLayerTest/0 (1028 ms total) -[----------] 3 tests from DummyDataLayerTest/1, where TypeParam = double -[ RUN ] DummyDataLayerTest/1.TestTwoTopConstant -[ OK ] DummyDataLayerTest/1.TestTwoTopConstant (0 ms) -[ RUN ] DummyDataLayerTest/1.TestOneTopConstant -[ OK ] DummyDataLayerTest/1.TestOneTopConstant (0 ms) -[ RUN ] DummyDataLayerTest/1.TestThreeTopConstantGaussianConstant -[ OK ] DummyDataLayerTest/1.TestThreeTopConstantGaussianConstant (0 ms) -[----------] 3 tests from DummyDataLayerTest/1 (0 ms total) +[----------] 2 tests from HDF5DataLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] HDF5DataLayerTest/1.TestRead +[ OK ] HDF5DataLayerTest/1.TestRead (14 ms) +[ RUN ] HDF5DataLayerTest/1.TestSkip +[ OK ] HDF5DataLayerTest/1.TestSkip (46 ms) +[----------] 2 tests from HDF5DataLayerTest/1 (60 ms total) -[----------] 2 tests from SoftmaxLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] SoftmaxLayerTest/0.TestGradient -[ OK ] SoftmaxLayerTest/0.TestGradient (303 ms) -[ RUN ] SoftmaxLayerTest/0.TestForward -[ OK ] SoftmaxLayerTest/0.TestForward (0 ms) -[----------] 2 tests from SoftmaxLayerTest/0 (303 ms total) +[----------] 15 tests from ConvolutionLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] ConvolutionLayerTest/1.Test1x1Convolution +[ OK ] ConvolutionLayerTest/1.Test1x1Convolution (1 ms) +[ RUN ] ConvolutionLayerTest/1.TestSimple3DConvolution +[ OK ] ConvolutionLayerTest/1.TestSimple3DConvolution (3 ms) +[ RUN ] ConvolutionLayerTest/1.TestSobelConvolution +[ OK ] ConvolutionLayerTest/1.TestSobelConvolution (0 ms) +[ RUN ] ConvolutionLayerTest/1.TestGradientGroup +[ OK ] ConvolutionLayerTest/1.TestGradientGroup (35 ms) +[ RUN ] ConvolutionLayerTest/1.TestDilated3DConvolution +[ OK ] ConvolutionLayerTest/1.TestDilated3DConvolution (16 ms) +[ RUN ] ConvolutionLayerTest/1.Test1x1Gradient +[ OK ] ConvolutionLayerTest/1.Test1x1Gradient (994 ms) +[ RUN ] ConvolutionLayerTest/1.TestSetup +[ OK ] ConvolutionLayerTest/1.TestSetup (0 ms) +[ RUN ] ConvolutionLayerTest/1.TestGradient +[ OK ] ConvolutionLayerTest/1.TestGradient (127 ms) +[ RUN ] ConvolutionLayerTest/1.TestDilatedConvolution +[ OK ] ConvolutionLayerTest/1.TestDilatedConvolution (3 ms) +[ RUN ] ConvolutionLayerTest/1.TestSimpleConvolutionGroup +[ OK ] ConvolutionLayerTest/1.TestSimpleConvolutionGroup (0 ms) +[ RUN ] ConvolutionLayerTest/1.TestNDAgainst2D +[ OK ] ConvolutionLayerTest/1.TestNDAgainst2D (734 ms) +[ RUN ] ConvolutionLayerTest/1.Test0DConvolution +[ OK ] ConvolutionLayerTest/1.Test0DConvolution (1 ms) +[ RUN ] ConvolutionLayerTest/1.TestDilatedGradient +[ OK ] ConvolutionLayerTest/1.TestDilatedGradient (33 ms) +[ RUN ] ConvolutionLayerTest/1.TestSimpleConvolution +[ OK ] ConvolutionLayerTest/1.TestSimpleConvolution (1 ms) +[ RUN ] ConvolutionLayerTest/1.TestGradient3D +[ OK ] ConvolutionLayerTest/1.TestGradient3D (1132 ms) +[----------] 15 tests from ConvolutionLayerTest/1 (3080 ms total) -[----------] 5 tests from MemoryDataLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] MemoryDataLayerTest/1.TestForward -[ OK ] MemoryDataLayerTest/1.TestForward (49 ms) -[ RUN ] MemoryDataLayerTest/1.AddMatVectorDefaultTransform -[ OK ] MemoryDataLayerTest/1.AddMatVectorDefaultTransform (1 ms) -[ RUN ] MemoryDataLayerTest/1.AddDatumVectorDefaultTransform -[ OK ] MemoryDataLayerTest/1.AddDatumVectorDefaultTransform (2 ms) -[ RUN ] MemoryDataLayerTest/1.TestSetBatchSize -[ OK ] MemoryDataLayerTest/1.TestSetBatchSize (2 ms) -[ RUN ] MemoryDataLayerTest/1.TestSetup -[ OK ] MemoryDataLayerTest/1.TestSetup (1 ms) -[----------] 5 tests from MemoryDataLayerTest/1 (56 ms total) +[----------] 27 tests from ReductionLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] ReductionLayerTest/0.TestMeanCoeffGradient +[ OK ] ReductionLayerTest/0.TestMeanCoeffGradient (1 ms) +[ RUN ] ReductionLayerTest/0.TestSum +[ OK ] ReductionLayerTest/0.TestSum (0 ms) +[ RUN ] ReductionLayerTest/0.TestMean +[ OK ] ReductionLayerTest/0.TestMean (0 ms) +[ RUN ] ReductionLayerTest/0.TestSumCoeffGradient +[ OK ] ReductionLayerTest/0.TestSumCoeffGradient (1 ms) +[ RUN ] ReductionLayerTest/0.TestSumOfSquaresCoeffAxis1Gradient +[ OK ] ReductionLayerTest/0.TestSumOfSquaresCoeffAxis1Gradient (2 ms) +[ RUN ] ReductionLayerTest/0.TestSetUpWithAxis2 +[ OK ] ReductionLayerTest/0.TestSetUpWithAxis2 (0 ms) +[ RUN ] ReductionLayerTest/0.TestSumCoeffAxis1 +[ OK ] ReductionLayerTest/0.TestSumCoeffAxis1 (0 ms) +[ RUN ] ReductionLayerTest/0.TestAbsSumCoeffAxis1Gradient +[ OK ] ReductionLayerTest/0.TestAbsSumCoeffAxis1Gradient (2 ms) +[ RUN ] ReductionLayerTest/0.TestAbsSum +[ OK ] ReductionLayerTest/0.TestAbsSum (1 ms) +[ RUN ] ReductionLayerTest/0.TestMeanCoeffGradientAxis1 +[ OK ] ReductionLayerTest/0.TestMeanCoeffGradientAxis1 (2 ms) +[ RUN ] ReductionLayerTest/0.TestSetUpWithAxis1 +[ OK ] ReductionLayerTest/0.TestSetUpWithAxis1 (0 ms) +[ RUN ] ReductionLayerTest/0.TestSumCoeffAxis1Gradient +[ OK ] ReductionLayerTest/0.TestSumCoeffAxis1Gradient (3 ms) +[ RUN ] ReductionLayerTest/0.TestSumCoeff +[ OK ] ReductionLayerTest/0.TestSumCoeff (0 ms) +[ RUN ] ReductionLayerTest/0.TestAbsSumCoeffAxis1 +[ OK ] ReductionLayerTest/0.TestAbsSumCoeffAxis1 (0 ms) +[ RUN ] ReductionLayerTest/0.TestSumOfSquaresCoeffAxis1 +[ OK ] ReductionLayerTest/0.TestSumOfSquaresCoeffAxis1 (0 ms) +[ RUN ] ReductionLayerTest/0.TestAbsSumCoeffGradient +[ OK ] ReductionLayerTest/0.TestAbsSumCoeffGradient (1 ms) +[ RUN ] ReductionLayerTest/0.TestSumOfSquaresCoeffGradient +[ OK ] ReductionLayerTest/0.TestSumOfSquaresCoeffGradient (1 ms) +[ RUN ] ReductionLayerTest/0.TestSumOfSquaresCoeff +[ OK ] ReductionLayerTest/0.TestSumOfSquaresCoeff (0 ms) +[ RUN ] ReductionLayerTest/0.TestSumOfSquares +[ OK ] ReductionLayerTest/0.TestSumOfSquares (0 ms) +[ RUN ] ReductionLayerTest/0.TestSetUp +[ OK ] ReductionLayerTest/0.TestSetUp (0 ms) +[ RUN ] ReductionLayerTest/0.TestMeanCoeffAxis1 +[ OK ] ReductionLayerTest/0.TestMeanCoeffAxis1 (0 ms) +[ RUN ] ReductionLayerTest/0.TestAbsSumGradient +[ OK ] ReductionLayerTest/0.TestAbsSumGradient (1 ms) +[ RUN ] ReductionLayerTest/0.TestMeanGradient +[ OK ] ReductionLayerTest/0.TestMeanGradient (2 ms) +[ RUN ] ReductionLayerTest/0.TestAbsSumCoeff +[ OK ] ReductionLayerTest/0.TestAbsSumCoeff (0 ms) +[ RUN ] ReductionLayerTest/0.TestSumGradient +[ OK ] ReductionLayerTest/0.TestSumGradient (1 ms) +[ RUN ] ReductionLayerTest/0.TestSumOfSquaresGradient +[ OK ] ReductionLayerTest/0.TestSumOfSquaresGradient (1 ms) +[ RUN ] ReductionLayerTest/0.TestMeanCoeff +[ OK ] ReductionLayerTest/0.TestMeanCoeff (0 ms) +[----------] 27 tests from ReductionLayerTest/0 (20 ms total) -[----------] 3 tests from TanHLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] TanHLayerTest/1.TestTanHOverflow -[ OK ] TanHLayerTest/1.TestTanHOverflow (0 ms) -[ RUN ] TanHLayerTest/1.TestTanHGradient -[ OK ] TanHLayerTest/1.TestTanHGradient (7 ms) -[ RUN ] TanHLayerTest/1.TestTanH -[ OK ] TanHLayerTest/1.TestTanH (0 ms) -[----------] 3 tests from TanHLayerTest/1 (8 ms total) +[----------] 7 tests from TileLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] TileLayerTest/0.TestTrivialGradient +[ OK ] TileLayerTest/0.TestTrivialGradient (121 ms) +[ RUN ] TileLayerTest/0.TestForwardChannels +[ OK ] TileLayerTest/0.TestForwardChannels (0 ms) +[ RUN ] TileLayerTest/0.TestForwardNum +[ OK ] TileLayerTest/0.TestForwardNum (0 ms) +[ RUN ] TileLayerTest/0.TestGradientNum +[ OK ] TileLayerTest/0.TestGradientNum (381 ms) +[ RUN ] TileLayerTest/0.TestTrivialSetup +[ OK ] TileLayerTest/0.TestTrivialSetup (0 ms) +[ RUN ] TileLayerTest/0.TestGradientChannels +[ OK ] TileLayerTest/0.TestGradientChannels (404 ms) +[ RUN ] TileLayerTest/0.TestSetup +[ OK ] TileLayerTest/0.TestSetup (0 ms) +[----------] 7 tests from TileLayerTest/0 (906 ms total) -[----------] 4 tests from ConstantFillerTest/0, where TypeParam = float -[ RUN ] ConstantFillerTest/0.TestFill -[ OK ] ConstantFillerTest/0.TestFill (0 ms) -[ RUN ] ConstantFillerTest/0.TestFill2D -[ OK ] ConstantFillerTest/0.TestFill2D (0 ms) -[ RUN ] ConstantFillerTest/0.TestFill1D -[ OK ] ConstantFillerTest/0.TestFill1D (0 ms) -[ RUN ] ConstantFillerTest/0.TestFill5D -[ OK ] ConstantFillerTest/0.TestFill5D (0 ms) -[----------] 4 tests from ConstantFillerTest/0 (0 ms total) +[----------] 8 tests from LRNLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] LRNLayerTest/1.TestForwardAcrossChannelsLargeRegion +[ OK ] LRNLayerTest/1.TestForwardAcrossChannelsLargeRegion (1 ms) +[ RUN ] LRNLayerTest/1.TestForwardWithinChannel +[ OK ] LRNLayerTest/1.TestForwardWithinChannel (0 ms) +[ RUN ] LRNLayerTest/1.TestGradientAcrossChannelsLargeRegion +[ OK ] LRNLayerTest/1.TestGradientAcrossChannelsLargeRegion (1746 ms) +[ RUN ] LRNLayerTest/1.TestGradientAcrossChannels +[ OK ] LRNLayerTest/1.TestGradientAcrossChannels (1418 ms) +[ RUN ] LRNLayerTest/1.TestSetupAcrossChannels +[ OK ] LRNLayerTest/1.TestSetupAcrossChannels (0 ms) +[ RUN ] LRNLayerTest/1.TestSetupWithinChannel +[ OK ] LRNLayerTest/1.TestSetupWithinChannel (0 ms) +[ RUN ] LRNLayerTest/1.TestForwardAcrossChannels +[ OK ] LRNLayerTest/1.TestForwardAcrossChannels (1 ms) +[ RUN ] LRNLayerTest/1.TestGradientWithinChannel +[ OK ] LRNLayerTest/1.TestGradientWithinChannel (1312 ms) +[----------] 8 tests from LRNLayerTest/1 (4478 ms total) -[----------] 11 tests from AdaDeltaSolverTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] AdaDeltaSolverTest/0.TestSnapshot -[ OK ] AdaDeltaSolverTest/0.TestSnapshot (33 ms) -[ RUN ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithMomentum -[ OK ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithMomentum (42 ms) -[ RUN ] AdaDeltaSolverTest/0.TestSnapshotShare -[ OK ] AdaDeltaSolverTest/0.TestSnapshotShare (40 ms) -[ RUN ] AdaDeltaSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum -[ OK ] AdaDeltaSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum (5 ms) -[ RUN ] AdaDeltaSolverTest/0.TestLeastSquaresUpdateWithMomentumMultiIter -[ OK ] AdaDeltaSolverTest/0.TestLeastSquaresUpdateWithMomentumMultiIter (108 ms) -[ RUN ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdate -[ OK ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdate (21 ms) -[ RUN ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithEverythingShare -[ OK ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithEverythingShare (115 ms) -[ RUN ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithHalfMomentum -[ OK ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithHalfMomentum (42 ms) -[ RUN ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithWeightDecay -[ OK ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithWeightDecay (19 ms) -[ RUN ] AdaDeltaSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare -[ OK ] AdaDeltaSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare (6 ms) -[ RUN ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithEverything -[ OK ] AdaDeltaSolverTest/0.TestAdaDeltaLeastSquaresUpdateWithEverything (99 ms) -[----------] 11 tests from AdaDeltaSolverTest/0 (530 ms total) +[----------] 20 tests from FilterNetTest +[ RUN ] FilterNetTest.TestFilterOutByMaxLevel +[ OK ] FilterNetTest.TestFilterOutByMaxLevel (1 ms) +[ RUN ] FilterNetTest.TestFilterInOutByIncludeMultiRule +[ OK ] FilterNetTest.TestFilterInOutByIncludeMultiRule (1 ms) +[ RUN ] FilterNetTest.TestNoFilter +[ OK ] FilterNetTest.TestNoFilter (0 ms) +[ RUN ] FilterNetTest.TestFilterInByStage2 +[ OK ] FilterNetTest.TestFilterInByStage2 (0 ms) +[ RUN ] FilterNetTest.TestFilterInByMaxLevel2 +[ OK ] FilterNetTest.TestFilterInByMaxLevel2 (1 ms) +[ RUN ] FilterNetTest.TestFilterInByMultipleStage2 +[ OK ] FilterNetTest.TestFilterInByMultipleStage2 (0 ms) +[ RUN ] FilterNetTest.TestFilterLeNetTrainTest +[ OK ] FilterNetTest.TestFilterLeNetTrainTest (2 ms) +[ RUN ] FilterNetTest.TestFilterOutByNotStage +[ OK ] FilterNetTest.TestFilterOutByNotStage (0 ms) +[ RUN ] FilterNetTest.TestFilterInByNotStage +[ OK ] FilterNetTest.TestFilterInByNotStage (0 ms) +[ RUN ] FilterNetTest.TestFilterOutByMinLevel +[ OK ] FilterNetTest.TestFilterOutByMinLevel (0 ms) +[ RUN ] FilterNetTest.TestFilterInByMinLevel +[ OK ] FilterNetTest.TestFilterInByMinLevel (0 ms) +[ RUN ] FilterNetTest.TestFilterInOutByExcludeMultiRule +[ OK ] FilterNetTest.TestFilterInOutByExcludeMultiRule (1 ms) +[ RUN ] FilterNetTest.TestFilterInByMultipleStage +[ OK ] FilterNetTest.TestFilterInByMultipleStage (0 ms) +[ RUN ] FilterNetTest.TestFilterInByMaxLevel +[ OK ] FilterNetTest.TestFilterInByMaxLevel (1 ms) +[ RUN ] FilterNetTest.TestFilterInByStage +[ OK ] FilterNetTest.TestFilterInByStage (0 ms) +[ RUN ] FilterNetTest.TestFilterInByIncludeMultiRule +[ OK ] FilterNetTest.TestFilterInByIncludeMultiRule (1 ms) +[ RUN ] FilterNetTest.TestFilterOutByStage2 +[ OK ] FilterNetTest.TestFilterOutByStage2 (0 ms) +[ RUN ] FilterNetTest.TestFilterInByMinLevel2 +[ OK ] FilterNetTest.TestFilterInByMinLevel2 (0 ms) +[ RUN ] FilterNetTest.TestFilterOutByMultipleStage +[ OK ] FilterNetTest.TestFilterOutByMultipleStage (1 ms) +[ RUN ] FilterNetTest.TestFilterOutByStage +[ OK ] FilterNetTest.TestFilterOutByStage (0 ms) +[----------] 20 tests from FilterNetTest (10 ms total) -[----------] 3 tests from BatchNormLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] BatchNormLayerTest/1.TestGradient -[ OK ] BatchNormLayerTest/1.TestGradient (298 ms) -[ RUN ] BatchNormLayerTest/1.TestForwardInplace -[ OK ] BatchNormLayerTest/1.TestForwardInplace (0 ms) -[ RUN ] BatchNormLayerTest/1.TestForward -[ OK ] BatchNormLayerTest/1.TestForward (1 ms) -[----------] 3 tests from BatchNormLayerTest/1 (299 ms total) +[----------] 1 test from SolverTypeUpgradeTest +[ RUN ] SolverTypeUpgradeTest.TestSimple +[ OK ] SolverTypeUpgradeTest.TestSimple (1 ms) +[----------] 1 test from SolverTypeUpgradeTest (1 ms total) -[----------] 3 tests from DummyDataLayerTest/0, where TypeParam = float -[ RUN ] DummyDataLayerTest/0.TestTwoTopConstant -[ OK ] DummyDataLayerTest/0.TestTwoTopConstant (0 ms) -[ RUN ] DummyDataLayerTest/0.TestOneTopConstant -[ OK ] DummyDataLayerTest/0.TestOneTopConstant (0 ms) -[ RUN ] DummyDataLayerTest/0.TestThreeTopConstantGaussianConstant -[ OK ] DummyDataLayerTest/0.TestThreeTopConstantGaussianConstant (0 ms) -[----------] 3 tests from DummyDataLayerTest/0 (0 ms total) +[----------] 10 tests from ConcatLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] ConcatLayerTest/0.TestSetupChannels +[ OK ] ConcatLayerTest/0.TestSetupChannels (0 ms) +[ RUN ] ConcatLayerTest/0.TestForwardTrivial +[ OK ] ConcatLayerTest/0.TestForwardTrivial (0 ms) +[ RUN ] ConcatLayerTest/0.TestGradientTrivial +[ OK ] ConcatLayerTest/0.TestGradientTrivial (5 ms) +[ RUN ] ConcatLayerTest/0.TestGradientChannelsBottomOneOnly +[ OK ] ConcatLayerTest/0.TestGradientChannelsBottomOneOnly (3 ms) +[ RUN ] ConcatLayerTest/0.TestForwardChannels +[ OK ] ConcatLayerTest/0.TestForwardChannels (1 ms) +[ RUN ] ConcatLayerTest/0.TestGradientNum +[ OK ] ConcatLayerTest/0.TestGradientNum (7 ms) +[ RUN ] ConcatLayerTest/0.TestSetupChannelsNegativeIndexing +[ OK ] ConcatLayerTest/0.TestSetupChannelsNegativeIndexing (0 ms) +[ RUN ] ConcatLayerTest/0.TestGradientChannels +[ OK ] ConcatLayerTest/0.TestGradientChannels (6 ms) +[ RUN ] ConcatLayerTest/0.TestSetupNum +[ OK ] ConcatLayerTest/0.TestSetupNum (0 ms) +[ RUN ] ConcatLayerTest/0.TestForwardNum +[ OK ] ConcatLayerTest/0.TestForwardNum (0 ms) +[----------] 10 tests from ConcatLayerTest/0 (22 ms total) -[----------] 4 tests from ConstantFillerTest/1, where TypeParam = double -[ RUN ] ConstantFillerTest/1.TestFill1D -[ OK ] ConstantFillerTest/1.TestFill1D (0 ms) -[ RUN ] ConstantFillerTest/1.TestFill -[ OK ] ConstantFillerTest/1.TestFill (0 ms) -[ RUN ] ConstantFillerTest/1.TestFill5D -[ OK ] ConstantFillerTest/1.TestFill5D (0 ms) -[ RUN ] ConstantFillerTest/1.TestFill2D -[ OK ] ConstantFillerTest/1.TestFill2D (0 ms) -[----------] 4 tests from ConstantFillerTest/1 (0 ms total) +[----------] 1 test from CPUStochasticPoolingLayerTest/1, where TypeParam = double +[ RUN ] CPUStochasticPoolingLayerTest/1.TestSetup +[ OK ] CPUStochasticPoolingLayerTest/1.TestSetup (0 ms) +[----------] 1 test from CPUStochasticPoolingLayerTest/1 (0 ms total) -[----------] 5 tests from SPPLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] SPPLayerTest/0.TestGradient -[ OK ] SPPLayerTest/0.TestGradient (3758 ms) -[ RUN ] SPPLayerTest/0.TestSetup -[ OK ] SPPLayerTest/0.TestSetup (0 ms) -[ RUN ] SPPLayerTest/0.TestForwardBackward -[ OK ] SPPLayerTest/0.TestForwardBackward (0 ms) -[ RUN ] SPPLayerTest/0.TestEqualOutputDims2 -[ OK ] SPPLayerTest/0.TestEqualOutputDims2 (0 ms) -[ RUN ] SPPLayerTest/0.TestEqualOutputDims -[ OK ] SPPLayerTest/0.TestEqualOutputDims (0 ms) -[----------] 5 tests from SPPLayerTest/0 (3758 ms total) +[----------] 2 tests from InfogainLossLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] InfogainLossLayerTest/0.TestGradient +[ OK ] InfogainLossLayerTest/0.TestGradient (3 ms) +[ RUN ] InfogainLossLayerTest/0.TestInfogainLoss +[ OK ] InfogainLossLayerTest/0.TestInfogainLoss (0 ms) +[----------] 2 tests from InfogainLossLayerTest/0 (3 ms total) + +[----------] 9 tests from AccuracyLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] AccuracyLayerTest/0.TestSetupTopK +[ OK ] AccuracyLayerTest/0.TestSetupTopK (0 ms) +[ RUN ] AccuracyLayerTest/0.TestSetupOutputPerClass +[ OK ] AccuracyLayerTest/0.TestSetupOutputPerClass (0 ms) +[ RUN ] AccuracyLayerTest/0.TestForward +[ OK ] AccuracyLayerTest/0.TestForward (1 ms) +[ RUN ] AccuracyLayerTest/0.TestForwardPerClass +[ OK ] AccuracyLayerTest/0.TestForwardPerClass (0 ms) +[ RUN ] AccuracyLayerTest/0.TestForwardTopK +[ OK ] AccuracyLayerTest/0.TestForwardTopK (6 ms) +[ RUN ] AccuracyLayerTest/0.TestForwardWithSpatialAxes +[ OK ] AccuracyLayerTest/0.TestForwardWithSpatialAxes (1 ms) +[ RUN ] AccuracyLayerTest/0.TestForwardPerClassWithIgnoreLabel +[ OK ] AccuracyLayerTest/0.TestForwardPerClassWithIgnoreLabel (0 ms) +[ RUN ] AccuracyLayerTest/0.TestSetup +[ OK ] AccuracyLayerTest/0.TestSetup (0 ms) +[ RUN ] AccuracyLayerTest/0.TestForwardIgnoreLabel +[ OK ] AccuracyLayerTest/0.TestForwardIgnoreLabel (0 ms) +[----------] 9 tests from AccuracyLayerTest/0 (11 ms total) + +[----------] 22 tests from ScaleLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] ScaleLayerTest/0.TestForwardEltwise +[ OK ] ScaleLayerTest/0.TestForwardEltwise (1 ms) +[ RUN ] ScaleLayerTest/0.TestForwardBroadcastEnd +[ OK ] ScaleLayerTest/0.TestForwardBroadcastEnd (0 ms) +[ RUN ] ScaleLayerTest/0.TestForwardBroadcastMiddleInPlace +[ OK ] ScaleLayerTest/0.TestForwardBroadcastMiddleInPlace (0 ms) +[ RUN ] ScaleLayerTest/0.TestGradientBroadcastEnd +[ OK ] ScaleLayerTest/0.TestGradientBroadcastEnd (328 ms) +[ RUN ] ScaleLayerTest/0.TestGradientEltwise +[ OK ] ScaleLayerTest/0.TestGradientEltwise (9 ms) +[ RUN ] ScaleLayerTest/0.TestGradientBroadcastMiddle +[ OK ] ScaleLayerTest/0.TestGradientBroadcastMiddle (175 ms) +[ RUN ] ScaleLayerTest/0.TestBackwardEltwiseInPlace +[ OK ] ScaleLayerTest/0.TestBackwardEltwiseInPlace (0 ms) +[ RUN ] ScaleLayerTest/0.TestForwardBroadcastMiddleWithParam +[ OK ] ScaleLayerTest/0.TestForwardBroadcastMiddleWithParam (0 ms) +[ RUN ] ScaleLayerTest/0.TestGradientBroadcastBegin +[ OK ] ScaleLayerTest/0.TestGradientBroadcastBegin (146 ms) +[ RUN ] ScaleLayerTest/0.TestForwardBroadcastMiddle +[ OK ] ScaleLayerTest/0.TestForwardBroadcastMiddle (0 ms) +[ RUN ] ScaleLayerTest/0.TestForwardBroadcastBegin +[ OK ] ScaleLayerTest/0.TestForwardBroadcastBegin (0 ms) +[ RUN ] ScaleLayerTest/0.TestForwardScale +[ OK ] ScaleLayerTest/0.TestForwardScale (0 ms) +[ RUN ] ScaleLayerTest/0.TestGradientScaleAxis2 +[ OK ] ScaleLayerTest/0.TestGradientScaleAxis2 (126 ms) +[ RUN ] ScaleLayerTest/0.TestGradientEltwiseWithParam +[ OK ] ScaleLayerTest/0.TestGradientEltwiseWithParam (572 ms) +[ RUN ] ScaleLayerTest/0.TestForwardBroadcastMiddleWithParamAndBias +[ OK ] ScaleLayerTest/0.TestForwardBroadcastMiddleWithParamAndBias (0 ms) +[ RUN ] ScaleLayerTest/0.TestForwardEltwiseWithParam +[ OK ] ScaleLayerTest/0.TestForwardEltwiseWithParam (0 ms) +[ RUN ] ScaleLayerTest/0.TestForwardEltwiseInPlace +[ OK ] ScaleLayerTest/0.TestForwardEltwiseInPlace (0 ms) +[ RUN ] ScaleLayerTest/0.TestGradientScale +[ OK ] ScaleLayerTest/0.TestGradientScale (130 ms) +[ RUN ] ScaleLayerTest/0.TestForwardScaleAxis2 +[ OK ] ScaleLayerTest/0.TestForwardScaleAxis2 (0 ms) +[ RUN ] ScaleLayerTest/0.TestGradientScaleAndBias +[ OK ] ScaleLayerTest/0.TestGradientScaleAndBias (174 ms) +[ RUN ] ScaleLayerTest/0.TestBackwardBroadcastMiddleInPlace +[ OK ] ScaleLayerTest/0.TestBackwardBroadcastMiddleInPlace (0 ms) +[ RUN ] ScaleLayerTest/0.TestGradientBroadcastMiddleWithParam +[ OK ] ScaleLayerTest/0.TestGradientBroadcastMiddleWithParam (176 ms) +[----------] 22 tests from ScaleLayerTest/0 (1837 ms total) + +[----------] 11 tests from PoolingLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] PoolingLayerTest/0.TestForwardAve +[ OK ] PoolingLayerTest/0.TestForwardAve (0 ms) +[ RUN ] PoolingLayerTest/0.TestGradientMax +[ OK ] PoolingLayerTest/0.TestGradientMax (1205 ms) +[ RUN ] PoolingLayerTest/0.TestForwardMax +[ OK ] PoolingLayerTest/0.TestForwardMax (1 ms) +[ RUN ] PoolingLayerTest/0.TestGradientMaxTopMask +[ OK ] PoolingLayerTest/0.TestGradientMaxTopMask (935 ms) +[ RUN ] PoolingLayerTest/0.TestGradientAve +[ OK ] PoolingLayerTest/0.TestGradientAve (372 ms) +[ RUN ] PoolingLayerTest/0.TestForwardMaxPadded +[ OK ] PoolingLayerTest/0.TestForwardMaxPadded (0 ms) +[ RUN ] PoolingLayerTest/0.TestForwardMaxTopMask +[ OK ] PoolingLayerTest/0.TestForwardMaxTopMask (1 ms) +[ RUN ] PoolingLayerTest/0.TestSetupGlobalPooling +[ OK ] PoolingLayerTest/0.TestSetupGlobalPooling (0 ms) +[ RUN ] PoolingLayerTest/0.TestGradientAvePadded +[ OK ] PoolingLayerTest/0.TestGradientAvePadded (1512 ms) +[ RUN ] PoolingLayerTest/0.TestSetup +[ OK ] PoolingLayerTest/0.TestSetup (0 ms) +[ RUN ] PoolingLayerTest/0.TestSetupPadded +[ OK ] PoolingLayerTest/0.TestSetupPadded (0 ms) +[----------] 11 tests from PoolingLayerTest/0 (4026 ms total) + +[----------] 1 test from SolverTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] SolverTest/0.TestInitTrainTestNets +[ OK ] SolverTest/0.TestInitTrainTestNets (3 ms) +[----------] 1 test from SolverTest/0 (3 ms total) [----------] 3 tests from MaxPoolingDropoutTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] MaxPoolingDropoutTest/1.TestForward -[ OK ] MaxPoolingDropoutTest/1.TestForward (1 ms) [ RUN ] MaxPoolingDropoutTest/1.TestSetup [ OK ] MaxPoolingDropoutTest/1.TestSetup (0 ms) +[ RUN ] MaxPoolingDropoutTest/1.TestForward +[ OK ] MaxPoolingDropoutTest/1.TestForward (0 ms) [ RUN ] MaxPoolingDropoutTest/1.TestBackward [ OK ] MaxPoolingDropoutTest/1.TestBackward (0 ms) -[----------] 3 tests from MaxPoolingDropoutTest/1 (1 ms total) +[----------] 3 tests from MaxPoolingDropoutTest/1 (0 ms total) -[----------] 12 tests from DataTransformTest/0, where TypeParam = float -[ RUN ] DataTransformTest/0.TestCropSize -[ OK ] DataTransformTest/0.TestCropSize (0 ms) -[ RUN ] DataTransformTest/0.TestCropTest -[ OK ] DataTransformTest/0.TestCropTest (0 ms) -[ RUN ] DataTransformTest/0.TestMeanValues -[ OK ] DataTransformTest/0.TestMeanValues (0 ms) -[ RUN ] DataTransformTest/0.TestMirrorTest -[ OK ] DataTransformTest/0.TestMirrorTest (0 ms) -[ RUN ] DataTransformTest/0.TestCropMirrorTest -[ OK ] DataTransformTest/0.TestCropMirrorTest (0 ms) -[ RUN ] DataTransformTest/0.TestCropMirrorTrain -[ OK ] DataTransformTest/0.TestCropMirrorTrain (0 ms) -[ RUN ] DataTransformTest/0.TestMirrorTrain -[ OK ] DataTransformTest/0.TestMirrorTrain (0 ms) -[ RUN ] DataTransformTest/0.TestMeanFile -[ OK ] DataTransformTest/0.TestMeanFile (0 ms) -[ RUN ] DataTransformTest/0.TestCropTrain -[ OK ] DataTransformTest/0.TestCropTrain (1 ms) -[ RUN ] DataTransformTest/0.TestMeanValue -[ OK ] DataTransformTest/0.TestMeanValue (0 ms) -[ RUN ] DataTransformTest/0.TestEmptyTransform -[ OK ] DataTransformTest/0.TestEmptyTransform (0 ms) -[ RUN ] DataTransformTest/0.TestEmptyTransformUniquePixels -[ OK ] DataTransformTest/0.TestEmptyTransformUniquePixels (0 ms) -[----------] 12 tests from DataTransformTest/0 (1 ms total) - -[----------] 4 tests from ContrastiveLossLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] ContrastiveLossLayerTest/0.TestForwardLegacy -[ OK ] ContrastiveLossLayerTest/0.TestForwardLegacy (0 ms) -[ RUN ] ContrastiveLossLayerTest/0.TestGradientLegacy -[ OK ] ContrastiveLossLayerTest/0.TestGradientLegacy (326 ms) -[ RUN ] ContrastiveLossLayerTest/0.TestForward -[ OK ] ContrastiveLossLayerTest/0.TestForward (1 ms) -[ RUN ] ContrastiveLossLayerTest/0.TestGradient -[ OK ] ContrastiveLossLayerTest/0.TestGradient (344 ms) -[----------] 4 tests from ContrastiveLossLayerTest/0 (671 ms total) - -[----------] 2 tests from InfogainLossLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] InfogainLossLayerTest/0.TestInfogainLoss -[ OK ] InfogainLossLayerTest/0.TestInfogainLoss (0 ms) -[ RUN ] InfogainLossLayerTest/0.TestGradient -[ OK ] InfogainLossLayerTest/0.TestGradient (3 ms) -[----------] 2 tests from InfogainLossLayerTest/0 (3 ms total) +[----------] 1 test from SolverFactoryTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] SolverFactoryTest/0.TestCreateSolver +[ OK ] SolverFactoryTest/0.TestCreateSolver (2 ms) +[----------] 1 test from SolverFactoryTest/0 (2 ms total) [----------] 4 tests from GaussianFillerTest/1, where TypeParam = double [ RUN ] GaussianFillerTest/1.TestFill2D [ OK ] GaussianFillerTest/1.TestFill2D (0 ms) -[ RUN ] GaussianFillerTest/1.TestFill5D -[ OK ] GaussianFillerTest/1.TestFill5D (0 ms) [ RUN ] GaussianFillerTest/1.TestFill1D [ OK ] GaussianFillerTest/1.TestFill1D (1 ms) [ RUN ] GaussianFillerTest/1.TestFill [ OK ] GaussianFillerTest/1.TestFill (0 ms) +[ RUN ] GaussianFillerTest/1.TestFill5D +[ OK ] GaussianFillerTest/1.TestFill5D (1 ms) [----------] 4 tests from GaussianFillerTest/1 (2 ms total) -[----------] 6 tests from RNNLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] RNNLayerTest/0.TestForward -[ OK ] RNNLayerTest/0.TestForward (9 ms) -[ RUN ] RNNLayerTest/0.TestGradientNonZeroContBufferSize2WithStaticInput -[ OK ] RNNLayerTest/0.TestGradientNonZeroContBufferSize2WithStaticInput (1645 ms) -[ RUN ] RNNLayerTest/0.TestGradientNonZeroCont -[ OK ] RNNLayerTest/0.TestGradientNonZeroCont (184 ms) -[ RUN ] RNNLayerTest/0.TestGradientNonZeroContBufferSize2 -[ OK ] RNNLayerTest/0.TestGradientNonZeroContBufferSize2 (372 ms) -[ RUN ] RNNLayerTest/0.TestGradient -[ OK ] RNNLayerTest/0.TestGradient (182 ms) -[ RUN ] RNNLayerTest/0.TestSetUp -[ OK ] RNNLayerTest/0.TestSetUp (2 ms) -[----------] 6 tests from RNNLayerTest/0 (2394 ms total) +[----------] 7 tests from CPUMathFunctionsTest/1, where TypeParam = double +[ RUN ] CPUMathFunctionsTest/1.TestSign +[ OK ] CPUMathFunctionsTest/1.TestSign (8 ms) +[ RUN ] CPUMathFunctionsTest/1.TestAsum +[ OK ] CPUMathFunctionsTest/1.TestAsum (6 ms) +[ RUN ] CPUMathFunctionsTest/1.TestFabs +[ OK ] CPUMathFunctionsTest/1.TestFabs (8 ms) +[ RUN ] CPUMathFunctionsTest/1.TestScale +[ OK ] CPUMathFunctionsTest/1.TestScale (7 ms) +[ RUN ] CPUMathFunctionsTest/1.TestNothing +[ OK ] CPUMathFunctionsTest/1.TestNothing (6 ms) +[ RUN ] CPUMathFunctionsTest/1.TestSgnbit +[ OK ] CPUMathFunctionsTest/1.TestSgnbit (7 ms) +[ RUN ] CPUMathFunctionsTest/1.TestCopy +[ OK ] CPUMathFunctionsTest/1.TestCopy (6 ms) +[----------] 7 tests from CPUMathFunctionsTest/1 (48 ms total) -[----------] 7 tests from TileLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] TileLayerTest/1.TestTrivialGradient -[ OK ] TileLayerTest/1.TestTrivialGradient (120 ms) -[ RUN ] TileLayerTest/1.TestGradientNum -[ OK ] TileLayerTest/1.TestGradientNum (383 ms) -[ RUN ] TileLayerTest/1.TestForwardChannels -[ OK ] TileLayerTest/1.TestForwardChannels (0 ms) -[ RUN ] TileLayerTest/1.TestGradientChannels -[ OK ] TileLayerTest/1.TestGradientChannels (411 ms) -[ RUN ] TileLayerTest/1.TestSetup -[ OK ] TileLayerTest/1.TestSetup (0 ms) -[ RUN ] TileLayerTest/1.TestForwardNum -[ OK ] TileLayerTest/1.TestForwardNum (0 ms) -[ RUN ] TileLayerTest/1.TestTrivialSetup -[ OK ] TileLayerTest/1.TestTrivialSetup (1 ms) -[----------] 7 tests from TileLayerTest/1 (915 ms total) +[----------] 3 tests from SplitLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] SplitLayerTest/0.Test +[ OK ] SplitLayerTest/0.Test (0 ms) +[ RUN ] SplitLayerTest/0.TestGradient +[ OK ] SplitLayerTest/0.TestGradient (9 ms) +[ RUN ] SplitLayerTest/0.TestSetup +[ OK ] SplitLayerTest/0.TestSetup (0 ms) +[----------] 3 tests from SplitLayerTest/0 (9 ms total) -[----------] 1 test from SolverTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] SolverTest/0.TestInitTrainTestNets -[ OK ] SolverTest/0.TestInitTrainTestNets (3 ms) -[----------] 1 test from SolverTest/0 (3 ms total) +[----------] 4 tests from UniformFillerTest/0, where TypeParam = float +[ RUN ] UniformFillerTest/0.TestFill1D +[ OK ] UniformFillerTest/0.TestFill1D (0 ms) +[ RUN ] UniformFillerTest/0.TestFill5D +[ OK ] UniformFillerTest/0.TestFill5D (0 ms) +[ RUN ] UniformFillerTest/0.TestFill2D +[ OK ] UniformFillerTest/0.TestFill2D (0 ms) +[ RUN ] UniformFillerTest/0.TestFill +[ OK ] UniformFillerTest/0.TestFill (0 ms) +[----------] 4 tests from UniformFillerTest/0 (0 ms total) -[----------] 11 tests from CropLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] CropLayerTest/1.TestCropAllOffset -[ OK ] CropLayerTest/1.TestCropAllOffset (0 ms) -[ RUN ] CropLayerTest/1.TestDimensionsCheck -[ OK ] CropLayerTest/1.TestDimensionsCheck (0 ms) -[ RUN ] CropLayerTest/1.TestCropHW -[ OK ] CropLayerTest/1.TestCropHW (0 ms) -[ RUN ] CropLayerTest/1.TestSetupShapeAll -[ OK ] CropLayerTest/1.TestSetupShapeAll (0 ms) -[ RUN ] CropLayerTest/1.TestSetupShapeNegativeIndexing -[ OK ] CropLayerTest/1.TestSetupShapeNegativeIndexing (0 ms) -[ RUN ] CropLayerTest/1.TestCrop5D -[ OK ] CropLayerTest/1.TestCrop5D (0 ms) -[ RUN ] CropLayerTest/1.TestSetupShapeDefault -[ OK ] CropLayerTest/1.TestSetupShapeDefault (0 ms) -[ RUN ] CropLayerTest/1.TestCropAll -[ OK ] CropLayerTest/1.TestCropAll (0 ms) -[ RUN ] CropLayerTest/1.TestCropHWGradient -[ OK ] CropLayerTest/1.TestCropHWGradient (690 ms) -[ RUN ] CropLayerTest/1.TestCropAllGradient -[ OK ] CropLayerTest/1.TestCropAllGradient (422 ms) -[ RUN ] CropLayerTest/1.TestCrop5DGradient -[ OK ] CropLayerTest/1.TestCrop5DGradient (2627 ms) -[----------] 11 tests from CropLayerTest/1 (3739 ms total) +[----------] 6 tests from XavierFillerTest/1, where TypeParam = double +[ RUN ] XavierFillerTest/1.TestFill2D +[ OK ] XavierFillerTest/1.TestFill2D (1 ms) +[ RUN ] XavierFillerTest/1.TestFill1D +[ OK ] XavierFillerTest/1.TestFill1D (0 ms) +[ RUN ] XavierFillerTest/1.TestFillFanIn +[ OK ] XavierFillerTest/1.TestFillFanIn (63 ms) +[ RUN ] XavierFillerTest/1.TestFill5D +[ OK ] XavierFillerTest/1.TestFill5D (0 ms) +[ RUN ] XavierFillerTest/1.TestFillAverage +[ OK ] XavierFillerTest/1.TestFillAverage (65 ms) +[ RUN ] XavierFillerTest/1.TestFillFanOut +[ OK ] XavierFillerTest/1.TestFillFanOut (64 ms) +[----------] 6 tests from XavierFillerTest/1 (193 ms total) -[----------] 1 test from MultinomialLogisticLossLayerTest/0, where TypeParam = float -[ RUN ] MultinomialLogisticLossLayerTest/0.TestGradientCPU -[ OK ] MultinomialLogisticLossLayerTest/0.TestGradientCPU (1 ms) -[----------] 1 test from MultinomialLogisticLossLayerTest/0 (1 ms total) +[----------] 2 tests from CommonTest +[ RUN ] CommonTest.TestBrewMode +[ OK ] CommonTest.TestBrewMode (0 ms) +[ RUN ] CommonTest.TestRandSeedCPU +[ OK ] CommonTest.TestRandSeedCPU (0 ms) +[----------] 2 tests from CommonTest (0 ms total) -[----------] 7 tests from CPUMathFunctionsTest/0, where TypeParam = float -[ RUN ] CPUMathFunctionsTest/0.TestSgnbit -[ OK ] CPUMathFunctionsTest/0.TestSgnbit (5 ms) -[ RUN ] CPUMathFunctionsTest/0.TestCopy -[ OK ] CPUMathFunctionsTest/0.TestCopy (4 ms) -[ RUN ] CPUMathFunctionsTest/0.TestAsum -[ OK ] CPUMathFunctionsTest/0.TestAsum (4 ms) -[ RUN ] CPUMathFunctionsTest/0.TestScale -[ OK ] CPUMathFunctionsTest/0.TestScale (5 ms) -[ RUN ] CPUMathFunctionsTest/0.TestSign -[ OK ] CPUMathFunctionsTest/0.TestSign (6 ms) -[ RUN ] CPUMathFunctionsTest/0.TestFabs -[ OK ] CPUMathFunctionsTest/0.TestFabs (6 ms) -[ RUN ] CPUMathFunctionsTest/0.TestNothing -[ OK ] CPUMathFunctionsTest/0.TestNothing (3 ms) -[----------] 7 tests from CPUMathFunctionsTest/0 (34 ms total) +[----------] 4 tests from UniformFillerTest/1, where TypeParam = double +[ RUN ] UniformFillerTest/1.TestFill2D +[ OK ] UniformFillerTest/1.TestFill2D (0 ms) +[ RUN ] UniformFillerTest/1.TestFill1D +[ OK ] UniformFillerTest/1.TestFill1D (0 ms) +[ RUN ] UniformFillerTest/1.TestFill +[ OK ] UniformFillerTest/1.TestFill (0 ms) +[ RUN ] UniformFillerTest/1.TestFill5D +[ OK ] UniformFillerTest/1.TestFill5D (0 ms) +[----------] 4 tests from UniformFillerTest/1 (0 ms total) -[----------] 5 tests from MemoryDataLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] MemoryDataLayerTest/0.TestForward -[ OK ] MemoryDataLayerTest/0.TestForward (50 ms) -[ RUN ] MemoryDataLayerTest/0.AddDatumVectorDefaultTransform -[ OK ] MemoryDataLayerTest/0.AddDatumVectorDefaultTransform (1 ms) -[ RUN ] MemoryDataLayerTest/0.AddMatVectorDefaultTransform -[ OK ] MemoryDataLayerTest/0.AddMatVectorDefaultTransform (2 ms) -[ RUN ] MemoryDataLayerTest/0.TestSetup -[ OK ] MemoryDataLayerTest/0.TestSetup (0 ms) -[ RUN ] MemoryDataLayerTest/0.TestSetBatchSize -[ OK ] MemoryDataLayerTest/0.TestSetBatchSize (3 ms) -[----------] 5 tests from MemoryDataLayerTest/0 (56 ms total) +[----------] 8 tests from Im2colLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] Im2colLayerTest/1.TestGradient +[ OK ] Im2colLayerTest/1.TestGradient (456 ms) +[ RUN ] Im2colLayerTest/1.TestDilatedGradient +[ OK ] Im2colLayerTest/1.TestDilatedGradient (1368 ms) +[ RUN ] Im2colLayerTest/1.TestSetup +[ OK ] Im2colLayerTest/1.TestSetup (0 ms) +[ RUN ] Im2colLayerTest/1.TestRect +[ OK ] Im2colLayerTest/1.TestRect (0 ms) +[ RUN ] Im2colLayerTest/1.TestRectGradient +[ OK ] Im2colLayerTest/1.TestRectGradient (376 ms) +[ RUN ] Im2colLayerTest/1.TestDilatedGradientForceND +[ OK ] Im2colLayerTest/1.TestDilatedGradientForceND (2559 ms) +[ RUN ] Im2colLayerTest/1.TestGradientForceND +[ OK ] Im2colLayerTest/1.TestGradientForceND (852 ms) +[ RUN ] Im2colLayerTest/1.TestForward +[ OK ] Im2colLayerTest/1.TestForward (0 ms) +[----------] 8 tests from Im2colLayerTest/1 (5611 ms total) -[----------] 7 tests from CPUMathFunctionsTest/1, where TypeParam = double -[ RUN ] CPUMathFunctionsTest/1.TestCopy -[ OK ] CPUMathFunctionsTest/1.TestCopy (6 ms) -[ RUN ] CPUMathFunctionsTest/1.TestFabs -[ OK ] CPUMathFunctionsTest/1.TestFabs (7 ms) -[ RUN ] CPUMathFunctionsTest/1.TestAsum -[ OK ] CPUMathFunctionsTest/1.TestAsum (5 ms) -[ RUN ] CPUMathFunctionsTest/1.TestScale -[ OK ] CPUMathFunctionsTest/1.TestScale (7 ms) -[ RUN ] CPUMathFunctionsTest/1.TestSgnbit -[ OK ] CPUMathFunctionsTest/1.TestSgnbit (8 ms) -[ RUN ] CPUMathFunctionsTest/1.TestNothing -[ OK ] CPUMathFunctionsTest/1.TestNothing (5 ms) -[ RUN ] CPUMathFunctionsTest/1.TestSign -[ OK ] CPUMathFunctionsTest/1.TestSign (7 ms) -[----------] 7 tests from CPUMathFunctionsTest/1 (47 ms total) +[----------] 12 tests from ReshapeLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] ReshapeLayerTest/1.TestInferenceOfUnspecifiedWithStartAxis +[ OK ] ReshapeLayerTest/1.TestInferenceOfUnspecifiedWithStartAxis (1 ms) +[ RUN ] ReshapeLayerTest/1.TestInferenceOfUnspecified +[ OK ] ReshapeLayerTest/1.TestInferenceOfUnspecified (0 ms) +[ RUN ] ReshapeLayerTest/1.TestInsertSingletonAxesMiddle +[ OK ] ReshapeLayerTest/1.TestInsertSingletonAxesMiddle (0 ms) +[ RUN ] ReshapeLayerTest/1.TestForward +[ OK ] ReshapeLayerTest/1.TestForward (0 ms) +[ RUN ] ReshapeLayerTest/1.TestCopyDimensions +[ OK ] ReshapeLayerTest/1.TestCopyDimensions (0 ms) +[ RUN ] ReshapeLayerTest/1.TestInsertSingletonAxesEnd +[ OK ] ReshapeLayerTest/1.TestInsertSingletonAxesEnd (0 ms) +[ RUN ] ReshapeLayerTest/1.TestFlattenValues +[ OK ] ReshapeLayerTest/1.TestFlattenValues (0 ms) +[ RUN ] ReshapeLayerTest/1.TestGradient +[ OK ] ReshapeLayerTest/1.TestGradient (4 ms) +[ RUN ] ReshapeLayerTest/1.TestInsertSingletonAxesStart +[ OK ] ReshapeLayerTest/1.TestInsertSingletonAxesStart (0 ms) +[ RUN ] ReshapeLayerTest/1.TestForwardAfterReshape +[ OK ] ReshapeLayerTest/1.TestForwardAfterReshape (0 ms) +[ RUN ] ReshapeLayerTest/1.TestFlattenMiddle +[ OK ] ReshapeLayerTest/1.TestFlattenMiddle (0 ms) +[ RUN ] ReshapeLayerTest/1.TestFlattenOutputSizes +[ OK ] ReshapeLayerTest/1.TestFlattenOutputSizes (0 ms) +[----------] 12 tests from ReshapeLayerTest/1 (5 ms total) [----------] 5 tests from DeconvolutionLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] DeconvolutionLayerTest/0.TestGradient3D +[ OK ] DeconvolutionLayerTest/0.TestGradient3D (405 ms) [ RUN ] DeconvolutionLayerTest/0.TestSimpleDeconvolution [ OK ] DeconvolutionLayerTest/0.TestSimpleDeconvolution (0 ms) -[ RUN ] DeconvolutionLayerTest/0.TestNDAgainst2D -[ OK ] DeconvolutionLayerTest/0.TestNDAgainst2D (1778 ms) -[ RUN ] DeconvolutionLayerTest/0.TestGradient -[ OK ] DeconvolutionLayerTest/0.TestGradient (1064 ms) -[ RUN ] DeconvolutionLayerTest/0.TestGradient3D -[ OK ] DeconvolutionLayerTest/0.TestGradient3D (391 ms) [ RUN ] DeconvolutionLayerTest/0.TestSetup [ OK ] DeconvolutionLayerTest/0.TestSetup (0 ms) -[----------] 5 tests from DeconvolutionLayerTest/0 (3233 ms total) - -[----------] 2 tests from HingeLossLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] HingeLossLayerTest/1.TestGradientL2 -[ OK ] HingeLossLayerTest/1.TestGradientL2 (1 ms) -[ RUN ] HingeLossLayerTest/1.TestGradientL1 -[ OK ] HingeLossLayerTest/1.TestGradientL1 (0 ms) -[----------] 2 tests from HingeLossLayerTest/1 (1 ms total) +[ RUN ] DeconvolutionLayerTest/0.TestNDAgainst2D +[ OK ] DeconvolutionLayerTest/0.TestNDAgainst2D (1806 ms) +[ RUN ] DeconvolutionLayerTest/0.TestGradient +[ OK ] DeconvolutionLayerTest/0.TestGradient (1112 ms) +[----------] 5 tests from DeconvolutionLayerTest/0 (3324 ms total) -[----------] 4 tests from PositiveUnitballFillerTest/0, where TypeParam = float -[ RUN ] PositiveUnitballFillerTest/0.TestFill -[ OK ] PositiveUnitballFillerTest/0.TestFill (0 ms) -[ RUN ] PositiveUnitballFillerTest/0.TestFill2D -[ OK ] PositiveUnitballFillerTest/0.TestFill2D (0 ms) -[ RUN ] PositiveUnitballFillerTest/0.TestFill1D -[ OK ] PositiveUnitballFillerTest/0.TestFill1D (0 ms) -[ RUN ] PositiveUnitballFillerTest/0.TestFill5D -[ OK ] PositiveUnitballFillerTest/0.TestFill5D (0 ms) -[----------] 4 tests from PositiveUnitballFillerTest/0 (0 ms total) +[----------] 3 tests from TanHLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] TanHLayerTest/0.TestTanH +[ OK ] TanHLayerTest/0.TestTanH (0 ms) +[ RUN ] TanHLayerTest/0.TestTanHGradient +[ OK ] TanHLayerTest/0.TestTanHGradient (7 ms) +[ RUN ] TanHLayerTest/0.TestTanHOverflow +[ OK ] TanHLayerTest/0.TestTanHOverflow (1 ms) +[----------] 3 tests from TanHLayerTest/0 (8 ms total) -[----------] 8 tests from AdamSolverTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] AdamSolverTest/1.TestAdamLeastSquaresUpdateWithWeightDecay -[ OK ] AdamSolverTest/1.TestAdamLeastSquaresUpdateWithWeightDecay (24 ms) -[ RUN ] AdamSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare -[ OK ] AdamSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare (8 ms) -[ RUN ] AdamSolverTest/1.TestAdamLeastSquaresUpdate -[ OK ] AdamSolverTest/1.TestAdamLeastSquaresUpdate (23 ms) -[ RUN ] AdamSolverTest/1.TestSnapshotShare -[ OK ] AdamSolverTest/1.TestSnapshotShare (49 ms) -[ RUN ] AdamSolverTest/1.TestSnapshot -[ OK ] AdamSolverTest/1.TestSnapshot (38 ms) -[ RUN ] AdamSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum -[ OK ] AdamSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum (6 ms) -[ RUN ] AdamSolverTest/1.TestAdamLeastSquaresUpdateWithEverythingShare -[ OK ] AdamSolverTest/1.TestAdamLeastSquaresUpdateWithEverythingShare (123 ms) -[ RUN ] AdamSolverTest/1.TestAdamLeastSquaresUpdateWithEverything -[ OK ] AdamSolverTest/1.TestAdamLeastSquaresUpdateWithEverything (116 ms) -[----------] 8 tests from AdamSolverTest/1 (388 ms total) +[----------] 3 tests from BatchNormLayerTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] BatchNormLayerTest/0.TestGradient +[ OK ] BatchNormLayerTest/0.TestGradient (271 ms) +[ RUN ] BatchNormLayerTest/0.TestForward +[ OK ] BatchNormLayerTest/0.TestForward (0 ms) +[ RUN ] BatchNormLayerTest/0.TestForwardInplace +[ OK ] BatchNormLayerTest/0.TestForwardInplace (0 ms) +[----------] 3 tests from BatchNormLayerTest/0 (271 ms total) -[----------] 4 tests from SoftmaxWithLossLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] SoftmaxWithLossLayerTest/0.TestForwardIgnoreLabel -[ OK ] SoftmaxWithLossLayerTest/0.TestForwardIgnoreLabel (0 ms) -[ RUN ] SoftmaxWithLossLayerTest/0.TestGradientUnnormalized -[ OK ] SoftmaxWithLossLayerTest/0.TestGradientUnnormalized (15 ms) -[ RUN ] SoftmaxWithLossLayerTest/0.TestGradientIgnoreLabel -[ OK ] SoftmaxWithLossLayerTest/0.TestGradientIgnoreLabel (16 ms) -[ RUN ] SoftmaxWithLossLayerTest/0.TestGradient -[ OK ] SoftmaxWithLossLayerTest/0.TestGradient (16 ms) -[----------] 4 tests from SoftmaxWithLossLayerTest/0 (48 ms total) +[----------] 8 tests from SliceLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] SliceLayerTest/1.TestSliceAcrossNum +[ OK ] SliceLayerTest/1.TestSliceAcrossNum (0 ms) +[ RUN ] SliceLayerTest/1.TestSetupNum +[ OK ] SliceLayerTest/1.TestSetupNum (0 ms) +[ RUN ] SliceLayerTest/1.TestSetupChannels +[ OK ] SliceLayerTest/1.TestSetupChannels (0 ms) +[ RUN ] SliceLayerTest/1.TestGradientTrivial +[ OK ] SliceLayerTest/1.TestGradientTrivial (17 ms) +[ RUN ] SliceLayerTest/1.TestSliceAcrossChannels +[ OK ] SliceLayerTest/1.TestSliceAcrossChannels (1 ms) +[ RUN ] SliceLayerTest/1.TestGradientAcrossChannels +[ OK ] SliceLayerTest/1.TestGradientAcrossChannels (67 ms) +[ RUN ] SliceLayerTest/1.TestGradientAcrossNum +[ OK ] SliceLayerTest/1.TestGradientAcrossNum (57 ms) +[ RUN ] SliceLayerTest/1.TestTrivialSlice +[ OK ] SliceLayerTest/1.TestTrivialSlice (0 ms) +[----------] 8 tests from SliceLayerTest/1 (142 ms total) -[----------] 3 tests from ThresholdLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] ThresholdLayerTest/1.Test -[ OK ] ThresholdLayerTest/1.Test (0 ms) -[ RUN ] ThresholdLayerTest/1.Test2 -[ OK ] ThresholdLayerTest/1.Test2 (0 ms) -[ RUN ] ThresholdLayerTest/1.TestSetup -[ OK ] ThresholdLayerTest/1.TestSetup (0 ms) -[----------] 3 tests from ThresholdLayerTest/1 (0 ms total) +[----------] 26 tests from IOTest +[ RUN ] IOTest.TestReadImageToCVMatGray +[ OK ] IOTest.TestReadImageToCVMatGray (4 ms) +[ RUN ] IOTest.TestDecodeDatumToCVMatNativeGray +[ OK ] IOTest.TestDecodeDatumToCVMatNativeGray (2 ms) +[ RUN ] IOTest.TestReadImageToDatumResizedGray +[ OK ] IOTest.TestReadImageToDatumResizedGray (6 ms) +[ RUN ] IOTest.TestReadImageToDatumGray +[ OK ] IOTest.TestReadImageToDatumGray (4 ms) +[ RUN ] IOTest.TestReadImageToDatumContentGray +[ OK ] IOTest.TestReadImageToDatumContentGray (10 ms) +[ RUN ] IOTest.TestDecodeDatumToCVMatNative +[ OK ] IOTest.TestDecodeDatumToCVMatNative (6 ms) +[ RUN ] IOTest.TestReadImageToDatumReferenceResized +[ OK ] IOTest.TestReadImageToDatumReferenceResized (15 ms) +[ RUN ] IOTest.TestDecodeDatumToCVMatContent +[ OK ] IOTest.TestDecodeDatumToCVMatContent (19 ms) +[ RUN ] IOTest.TestReadImageToDatumResizedSquare +[ OK ] IOTest.TestReadImageToDatumResizedSquare (9 ms) +[ RUN ] IOTest.TestDecodeDatumToCVMat +[ OK ] IOTest.TestDecodeDatumToCVMat (10 ms) +[ RUN ] IOTest.TestCVMatToDatumContent +[ OK ] IOTest.TestCVMatToDatumContent (17 ms) +[ RUN ] IOTest.TestCVMatToDatumReference +[ OK ] IOTest.TestCVMatToDatumReference (21 ms) +[ RUN ] IOTest.TestReadImageToDatum +[ OK ] IOTest.TestReadImageToDatum (8 ms) +[ RUN ] IOTest.TestReadFileToDatum +[ OK ] IOTest.TestReadFileToDatum (0 ms) +[ RUN ] IOTest.TestReadImageToDatumContent +[ OK ] IOTest.TestReadImageToDatumContent (18 ms) +[ RUN ] IOTest.TestReadImageToDatumResized +[ OK ] IOTest.TestReadImageToDatumResized (8 ms) +[ RUN ] IOTest.TestReadImageToCVMat +[ OK ] IOTest.TestReadImageToCVMat (7 ms) +[ RUN ] IOTest.TestReadImageToDatumReference +[ OK ] IOTest.TestReadImageToDatumReference (19 ms) +[ RUN ] IOTest.TestReadImageToCVMatResized +[ OK ] IOTest.TestReadImageToCVMatResized (7 ms) +[ RUN ] IOTest.TestDecodeDatumNativeGray +[ OK ] IOTest.TestDecodeDatumNativeGray (9 ms) +[ RUN ] IOTest.TestDecodeDatum +[ OK ] IOTest.TestDecodeDatum (20 ms) +[ RUN ] IOTest.TestReadImageToCVMatResizedGray +[ OK ] IOTest.TestReadImageToCVMatResizedGray (6 ms) +[ RUN ] IOTest.TestReadImageToCVMatResizedSquare +[ OK ] IOTest.TestReadImageToCVMatResizedSquare (8 ms) +[ RUN ] IOTest.TestDecodeDatumToCVMatContentNative +[ OK ] IOTest.TestDecodeDatumToCVMatContentNative (19 ms) +[ RUN ] IOTest.TestCVMatToDatum +[ OK ] IOTest.TestCVMatToDatum (8 ms) +[ RUN ] IOTest.TestDecodeDatumNative +[ OK ] IOTest.TestDecodeDatumNative (20 ms) +[----------] 26 tests from IOTest (282 ms total) -[----------] 5 tests from BenchmarkTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] BenchmarkTest/1.TestTimerConstructor -[ OK ] BenchmarkTest/1.TestTimerConstructor (0 ms) -[ RUN ] BenchmarkTest/1.TestTimerSeconds -[ OK ] BenchmarkTest/1.TestTimerSeconds (300 ms) -[ RUN ] BenchmarkTest/1.TestTimerStop -[ OK ] BenchmarkTest/1.TestTimerStop (0 ms) -[ RUN ] BenchmarkTest/1.TestTimerStart -[ OK ] BenchmarkTest/1.TestTimerStart (0 ms) -[ RUN ] BenchmarkTest/1.TestTimerMilliSeconds -[ OK ] BenchmarkTest/1.TestTimerMilliSeconds (300 ms) -[----------] 5 tests from BenchmarkTest/1 (601 ms total) +[----------] 26 tests from NetTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] NetTest/1.TestHasBlob +[ OK ] NetTest/1.TestHasBlob (2 ms) +[ RUN ] NetTest/1.TestBottomNeedBackward +[ OK ] NetTest/1.TestBottomNeedBackward (2 ms) +[ RUN ] NetTest/1.TestHasLayer +[ OK ] NetTest/1.TestHasLayer (2 ms) +[ RUN ] NetTest/1.TestLossWeight +[ OK ] NetTest/1.TestLossWeight (8 ms) +[ RUN ] NetTest/1.TestAllInOneNetVal +[ OK ] NetTest/1.TestAllInOneNetVal (2 ms) +[ RUN ] NetTest/1.TestBottomNeedBackwardForce +[ OK ] NetTest/1.TestBottomNeedBackwardForce (2 ms) +[ RUN ] NetTest/1.TestUnsharedWeightsDataNet +[ OK ] NetTest/1.TestUnsharedWeightsDataNet (1 ms) +[ RUN ] NetTest/1.TestAllInOneNetTrain +[ OK ] NetTest/1.TestAllInOneNetTrain (1 ms) +[ RUN ] NetTest/1.TestSharedWeightsDiffNet +[ OK ] NetTest/1.TestSharedWeightsDiffNet (1 ms) +[ RUN ] NetTest/1.TestLossWeightMidNet +[ OK ] NetTest/1.TestLossWeightMidNet (10 ms) +[ RUN ] NetTest/1.TestSharedWeightsResume +[ OK ] NetTest/1.TestSharedWeightsResume (2 ms) +[ RUN ] NetTest/1.TestParamPropagateDown +[ OK ] NetTest/1.TestParamPropagateDown (5 ms) +[ RUN ] NetTest/1.TestComboLossWeight +[ OK ] NetTest/1.TestComboLossWeight (7 ms) +[ RUN ] NetTest/1.TestUnsharedWeightsDiffNet +[ OK ] NetTest/1.TestUnsharedWeightsDiffNet (1 ms) +[ RUN ] NetTest/1.TestAllInOneNetDeploy +[ OK ] NetTest/1.TestAllInOneNetDeploy (1 ms) +[ RUN ] NetTest/1.TestSharedWeightsUpdate +[ OK ] NetTest/1.TestSharedWeightsUpdate (2 ms) +[ RUN ] NetTest/1.TestFromTo +[ OK ] NetTest/1.TestFromTo (5 ms) +[ RUN ] NetTest/1.TestForcePropagateDown +[ OK ] NetTest/1.TestForcePropagateDown (1 ms) +[ RUN ] NetTest/1.TestReshape +[ OK ] NetTest/1.TestReshape (2 ms) +[ RUN ] NetTest/1.TestBackwardWithAccuracyLayer +[ OK ] NetTest/1.TestBackwardWithAccuracyLayer (4 ms) +[ RUN ] NetTest/1.TestSharedWeightsDataNet +[ OK ] NetTest/1.TestSharedWeightsDataNet (1 ms) +[ RUN ] NetTest/1.TestGetBlob +[ OK ] NetTest/1.TestGetBlob (2 ms) +[ RUN ] NetTest/1.TestSkipPropagateDown +[ OK ] NetTest/1.TestSkipPropagateDown (3 ms) +[ RUN ] NetTest/1.TestGetLayerByName +[ OK ] NetTest/1.TestGetLayerByName (2 ms) +[ RUN ] NetTest/1.TestBottomNeedBackwardEuclideanForce +[ OK ] NetTest/1.TestBottomNeedBackwardEuclideanForce (0 ms) +[ RUN ] NetTest/1.TestBottomNeedBackwardTricky +[ OK ] NetTest/1.TestBottomNeedBackwardTricky (2 ms) +[----------] 26 tests from NetTest/1 (71 ms total) -[----------] 9 tests from AdaGradSolverTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] AdaGradSolverTest/1.TestSnapshotShare -[ OK ] AdaGradSolverTest/1.TestSnapshotShare (54 ms) -[ RUN ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdateWithEverythingShare -[ OK ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdateWithEverythingShare (137 ms) -[ RUN ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdateLROneHundredth -[ OK ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdateLROneHundredth (22 ms) -[ RUN ] AdaGradSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare -[ OK ] AdaGradSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare (8 ms) -[ RUN ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdate -[ OK ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdate (26 ms) -[ RUN ] AdaGradSolverTest/1.TestSnapshot -[ OK ] AdaGradSolverTest/1.TestSnapshot (39 ms) -[ RUN ] AdaGradSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum -[ OK ] AdaGradSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum (6 ms) -[ RUN ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdateWithWeightDecay -[ OK ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdateWithWeightDecay (23 ms) -[ RUN ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdateWithEverything -[ OK ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdateWithEverything (120 ms) -[----------] 9 tests from AdaGradSolverTest/1 (435 ms total) +[----------] 5 tests from BenchmarkTest/0, where TypeParam = caffe::CPUDevice +[ RUN ] BenchmarkTest/0.TestTimerStart +[ OK ] BenchmarkTest/0.TestTimerStart (0 ms) +[ RUN ] BenchmarkTest/0.TestTimerStop +[ OK ] BenchmarkTest/0.TestTimerStop (0 ms) +[ RUN ] BenchmarkTest/0.TestTimerMilliSeconds +[ OK ] BenchmarkTest/0.TestTimerMilliSeconds (301 ms) +[ RUN ] BenchmarkTest/0.TestTimerConstructor +[ OK ] BenchmarkTest/0.TestTimerConstructor (0 ms) +[ RUN ] BenchmarkTest/0.TestTimerSeconds +[ OK ] BenchmarkTest/0.TestTimerSeconds (300 ms) +[----------] 5 tests from BenchmarkTest/0 (601 ms total) -[----------] 8 tests from AdamSolverTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] AdamSolverTest/0.TestAdamLeastSquaresUpdate -[ OK ] AdamSolverTest/0.TestAdamLeastSquaresUpdate (22 ms) -[ RUN ] AdamSolverTest/0.TestAdamLeastSquaresUpdateWithEverything -[ OK ] AdamSolverTest/0.TestAdamLeastSquaresUpdateWithEverything (108 ms) -[ RUN ] AdamSolverTest/0.TestAdamLeastSquaresUpdateWithWeightDecay -[ OK ] AdamSolverTest/0.TestAdamLeastSquaresUpdateWithWeightDecay (21 ms) -[ RUN ] AdamSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum -[ OK ] AdamSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum (5 ms) -[ RUN ] AdamSolverTest/0.TestSnapshot -[ OK ] AdamSolverTest/0.TestSnapshot (34 ms) -[ RUN ] AdamSolverTest/0.TestSnapshotShare -[ OK ] AdamSolverTest/0.TestSnapshotShare (48 ms) -[ RUN ] AdamSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare -[ OK ] AdamSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare (7 ms) -[ RUN ] AdamSolverTest/0.TestAdamLeastSquaresUpdateWithEverythingShare -[ OK ] AdamSolverTest/0.TestAdamLeastSquaresUpdateWithEverythingShare (117 ms) -[----------] 8 tests from AdamSolverTest/0 (363 ms total) +[----------] 1 test from MultinomialLogisticLossLayerTest/1, where TypeParam = double +[ RUN ] MultinomialLogisticLossLayerTest/1.TestGradientCPU +[ OK ] MultinomialLogisticLossLayerTest/1.TestGradientCPU (1 ms) +[----------] 1 test from MultinomialLogisticLossLayerTest/1 (1 ms total) -[----------] 2 tests from HDF5DataLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] HDF5DataLayerTest/1.TestSkip -[ OK ] HDF5DataLayerTest/1.TestSkip (46 ms) -[ RUN ] HDF5DataLayerTest/1.TestRead -[ OK ] HDF5DataLayerTest/1.TestRead (13 ms) -[----------] 2 tests from HDF5DataLayerTest/1 (59 ms total) +[----------] 11 tests from RandomNumberGeneratorTest/0, where TypeParam = float +[ RUN ] RandomNumberGeneratorTest/0.TestRngGaussianTimesGaussian +[ OK ] RandomNumberGeneratorTest/0.TestRngGaussianTimesGaussian (0 ms) +[ RUN ] RandomNumberGeneratorTest/0.TestRngUniform2 +[ OK ] RandomNumberGeneratorTest/0.TestRngUniform2 (1 ms) +[ RUN ] RandomNumberGeneratorTest/0.TestRngBernoulli2 +[ OK ] RandomNumberGeneratorTest/0.TestRngBernoulli2 (0 ms) +[ RUN ] RandomNumberGeneratorTest/0.TestRngBernoulliTimesBernoulli +[ OK ] RandomNumberGeneratorTest/0.TestRngBernoulliTimesBernoulli (0 ms) +[ RUN ] RandomNumberGeneratorTest/0.TestRngGaussian2 +[ OK ] RandomNumberGeneratorTest/0.TestRngGaussian2 (1 ms) +[ RUN ] RandomNumberGeneratorTest/0.TestRngUniformTimesBernoulli +[ OK ] RandomNumberGeneratorTest/0.TestRngUniformTimesBernoulli (0 ms) +[ RUN ] RandomNumberGeneratorTest/0.TestRngUniform +[ OK ] RandomNumberGeneratorTest/0.TestRngUniform (0 ms) +[ RUN ] RandomNumberGeneratorTest/0.TestRngBernoulli +[ OK ] RandomNumberGeneratorTest/0.TestRngBernoulli (1 ms) +[ RUN ] RandomNumberGeneratorTest/0.TestRngUniformTimesUniform +[ OK ] RandomNumberGeneratorTest/0.TestRngUniformTimesUniform (0 ms) +[ RUN ] RandomNumberGeneratorTest/0.TestRngGaussianTimesBernoulli +[ OK ] RandomNumberGeneratorTest/0.TestRngGaussianTimesBernoulli (1 ms) +[ RUN ] RandomNumberGeneratorTest/0.TestRngGaussian +[ OK ] RandomNumberGeneratorTest/0.TestRngGaussian (0 ms) +[----------] 11 tests from RandomNumberGeneratorTest/0 (4 ms total) -[----------] 3 tests from ThresholdLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] ThresholdLayerTest/0.Test2 -[ OK ] ThresholdLayerTest/0.Test2 (0 ms) -[ RUN ] ThresholdLayerTest/0.TestSetup -[ OK ] ThresholdLayerTest/0.TestSetup (0 ms) -[ RUN ] ThresholdLayerTest/0.Test -[ OK ] ThresholdLayerTest/0.Test (0 ms) -[----------] 3 tests from ThresholdLayerTest/0 (0 ms total) +[----------] 6 tests from RNNLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] RNNLayerTest/1.TestGradientNonZeroContBufferSize2 +[ OK ] RNNLayerTest/1.TestGradientNonZeroContBufferSize2 (442 ms) +[ RUN ] RNNLayerTest/1.TestSetUp +[ OK ] RNNLayerTest/1.TestSetUp (3 ms) +[ RUN ] RNNLayerTest/1.TestGradientNonZeroCont +[ OK ] RNNLayerTest/1.TestGradientNonZeroCont (242 ms) +[ RUN ] RNNLayerTest/1.TestGradientNonZeroContBufferSize2WithStaticInput +[ OK ] RNNLayerTest/1.TestGradientNonZeroContBufferSize2WithStaticInput (1882 ms) +[ RUN ] RNNLayerTest/1.TestForward +[ OK ] RNNLayerTest/1.TestForward (10 ms) +[ RUN ] RNNLayerTest/1.TestGradient +[ OK ] RNNLayerTest/1.TestGradient (199 ms) +[----------] 6 tests from RNNLayerTest/1 (2778 ms total) [----------] 6 tests from FlattenLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] FlattenLayerTest/0.TestSetupWithEndAxis -[ OK ] FlattenLayerTest/0.TestSetupWithEndAxis (0 ms) -[ RUN ] FlattenLayerTest/0.TestForward -[ OK ] FlattenLayerTest/0.TestForward (0 ms) +[ RUN ] FlattenLayerTest/0.TestSetupWithAxis +[ OK ] FlattenLayerTest/0.TestSetupWithAxis (0 ms) [ RUN ] FlattenLayerTest/0.TestSetup [ OK ] FlattenLayerTest/0.TestSetup (0 ms) +[ RUN ] FlattenLayerTest/0.TestForward +[ OK ] FlattenLayerTest/0.TestForward (0 ms) +[ RUN ] FlattenLayerTest/0.TestGradient +[ OK ] FlattenLayerTest/0.TestGradient (5 ms) [ RUN ] FlattenLayerTest/0.TestSetupWithStartAndEndAxis [ OK ] FlattenLayerTest/0.TestSetupWithStartAndEndAxis (0 ms) -[ RUN ] FlattenLayerTest/0.TestGradient -[ OK ] FlattenLayerTest/0.TestGradient (4 ms) -[ RUN ] FlattenLayerTest/0.TestSetupWithAxis -[ OK ] FlattenLayerTest/0.TestSetupWithAxis (0 ms) +[ RUN ] FlattenLayerTest/0.TestSetupWithEndAxis +[ OK ] FlattenLayerTest/0.TestSetupWithEndAxis (0 ms) [----------] 6 tests from FlattenLayerTest/0 (5 ms total) -[----------] 4 tests from GaussianFillerTest/0, where TypeParam = float -[ RUN ] GaussianFillerTest/0.TestFill -[ OK ] GaussianFillerTest/0.TestFill (0 ms) -[ RUN ] GaussianFillerTest/0.TestFill5D -[ OK ] GaussianFillerTest/0.TestFill5D (1 ms) -[ RUN ] GaussianFillerTest/0.TestFill2D -[ OK ] GaussianFillerTest/0.TestFill2D (0 ms) -[ RUN ] GaussianFillerTest/0.TestFill1D -[ OK ] GaussianFillerTest/0.TestFill1D (0 ms) -[----------] 4 tests from GaussianFillerTest/0 (1 ms total) - -[----------] 3 tests from BlobMathTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] BlobMathTest/1.TestSumOfSquares -[ OK ] BlobMathTest/1.TestSumOfSquares (0 ms) -[ RUN ] BlobMathTest/1.TestScaleData -[ OK ] BlobMathTest/1.TestScaleData (0 ms) -[ RUN ] BlobMathTest/1.TestAsum -[ OK ] BlobMathTest/1.TestAsum (0 ms) -[----------] 3 tests from BlobMathTest/1 (0 ms total) - -[----------] 11 tests from PoolingLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] PoolingLayerTest/1.TestForwardMaxTopMask -[ OK ] PoolingLayerTest/1.TestForwardMaxTopMask (1 ms) -[ RUN ] PoolingLayerTest/1.TestGradientAvePadded -[ OK ] PoolingLayerTest/1.TestGradientAvePadded (1538 ms) -[ RUN ] PoolingLayerTest/1.TestForwardMaxPadded -[ OK ] PoolingLayerTest/1.TestForwardMaxPadded (0 ms) -[ RUN ] PoolingLayerTest/1.TestGradientAve -[ OK ] PoolingLayerTest/1.TestGradientAve (375 ms) -[ RUN ] PoolingLayerTest/1.TestSetupGlobalPooling -[ OK ] PoolingLayerTest/1.TestSetupGlobalPooling (0 ms) -[ RUN ] PoolingLayerTest/1.TestGradientMaxTopMask -[ OK ] PoolingLayerTest/1.TestGradientMaxTopMask (940 ms) -[ RUN ] PoolingLayerTest/1.TestSetup -[ OK ] PoolingLayerTest/1.TestSetup (0 ms) -[ RUN ] PoolingLayerTest/1.TestGradientMax -[ OK ] PoolingLayerTest/1.TestGradientMax (1225 ms) -[ RUN ] PoolingLayerTest/1.TestForwardAve -[ OK ] PoolingLayerTest/1.TestForwardAve (0 ms) -[ RUN ] PoolingLayerTest/1.TestSetupPadded -[ OK ] PoolingLayerTest/1.TestSetupPadded (1 ms) -[ RUN ] PoolingLayerTest/1.TestForwardMax -[ OK ] PoolingLayerTest/1.TestForwardMax (0 ms) -[----------] 11 tests from PoolingLayerTest/1 (4080 ms total) - -[----------] 12 tests from ArgMaxLayerTest/1, where TypeParam = double -[ RUN ] ArgMaxLayerTest/1.TestSetup -[ OK ] ArgMaxLayerTest/1.TestSetup (1 ms) -[ RUN ] ArgMaxLayerTest/1.TestSetupMaxVal -[ OK ] ArgMaxLayerTest/1.TestSetupMaxVal (2 ms) -[ RUN ] ArgMaxLayerTest/1.TestCPUMaxVal -[ OK ] ArgMaxLayerTest/1.TestCPUMaxVal (2 ms) -[ RUN ] ArgMaxLayerTest/1.TestCPUAxis -[ OK ] ArgMaxLayerTest/1.TestCPUAxis (14 ms) -[ RUN ] ArgMaxLayerTest/1.TestSetupAxis -[ OK ] ArgMaxLayerTest/1.TestSetupAxis (1 ms) -[ RUN ] ArgMaxLayerTest/1.TestCPUMaxValTopK -[ OK ] ArgMaxLayerTest/1.TestCPUMaxValTopK (2 ms) -[ RUN ] ArgMaxLayerTest/1.TestCPUAxisTopK -[ OK ] ArgMaxLayerTest/1.TestCPUAxisTopK (55 ms) -[ RUN ] ArgMaxLayerTest/1.TestCPUTopK -[ OK ] ArgMaxLayerTest/1.TestCPUTopK (2 ms) -[ RUN ] ArgMaxLayerTest/1.TestCPUAxisMaxValTopK -[ OK ] ArgMaxLayerTest/1.TestCPUAxisMaxValTopK (51 ms) -[ RUN ] ArgMaxLayerTest/1.TestCPU -[ OK ] ArgMaxLayerTest/1.TestCPU (2 ms) -[ RUN ] ArgMaxLayerTest/1.TestSetupAxisMaxVal -[ OK ] ArgMaxLayerTest/1.TestSetupAxisMaxVal (2 ms) -[ RUN ] ArgMaxLayerTest/1.TestSetupAxisNegativeIndexing -[ OK ] ArgMaxLayerTest/1.TestSetupAxisNegativeIndexing (1 ms) -[----------] 12 tests from ArgMaxLayerTest/1 (135 ms total) - -[----------] 3 tests from MaxPoolingDropoutTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] MaxPoolingDropoutTest/0.TestForward -[ OK ] MaxPoolingDropoutTest/0.TestForward (0 ms) -[ RUN ] MaxPoolingDropoutTest/0.TestSetup -[ OK ] MaxPoolingDropoutTest/0.TestSetup (1 ms) -[ RUN ] MaxPoolingDropoutTest/0.TestBackward -[ OK ] MaxPoolingDropoutTest/0.TestBackward (0 ms) -[----------] 3 tests from MaxPoolingDropoutTest/0 (1 ms total) - -[----------] 6 tests from RNNLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] RNNLayerTest/1.TestGradient -[ OK ] RNNLayerTest/1.TestGradient (180 ms) -[ RUN ] RNNLayerTest/1.TestForward -[ OK ] RNNLayerTest/1.TestForward (10 ms) -[ RUN ] RNNLayerTest/1.TestSetUp -[ OK ] RNNLayerTest/1.TestSetUp (2 ms) -[ RUN ] RNNLayerTest/1.TestGradientNonZeroContBufferSize2WithStaticInput -[ OK ] RNNLayerTest/1.TestGradientNonZeroContBufferSize2WithStaticInput (1722 ms) -[ RUN ] RNNLayerTest/1.TestGradientNonZeroContBufferSize2 -[ OK ] RNNLayerTest/1.TestGradientNonZeroContBufferSize2 (370 ms) -[ RUN ] RNNLayerTest/1.TestGradientNonZeroCont -[ OK ] RNNLayerTest/1.TestGradientNonZeroCont (180 ms) -[----------] 6 tests from RNNLayerTest/1 (2464 ms total) - -[----------] 6 tests from MVNLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] MVNLayerTest/0.TestForward -[ OK ] MVNLayerTest/0.TestForward (0 ms) -[ RUN ] MVNLayerTest/0.TestGradientMeanOnly -[ OK ] MVNLayerTest/0.TestGradientMeanOnly (165 ms) -[ RUN ] MVNLayerTest/0.TestGradient -[ OK ] MVNLayerTest/0.TestGradient (430 ms) -[ RUN ] MVNLayerTest/0.TestGradientAcrossChannels -[ OK ] MVNLayerTest/0.TestGradientAcrossChannels (412 ms) -[ RUN ] MVNLayerTest/0.TestForwardAcrossChannels -[ OK ] MVNLayerTest/0.TestForwardAcrossChannels (0 ms) -[ RUN ] MVNLayerTest/0.TestForwardMeanOnly -[ OK ] MVNLayerTest/0.TestForwardMeanOnly (0 ms) -[----------] 6 tests from MVNLayerTest/0 (1008 ms total) - -[----------] 20 tests from BiasLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] BiasLayerTest/0.TestForwardBroadcastEnd -[ OK ] BiasLayerTest/0.TestForwardBroadcastEnd (0 ms) -[ RUN ] BiasLayerTest/0.TestForwardEltwiseWithParam -[ OK ] BiasLayerTest/0.TestForwardEltwiseWithParam (0 ms) -[ RUN ] BiasLayerTest/0.TestForwardBias -[ OK ] BiasLayerTest/0.TestForwardBias (0 ms) -[ RUN ] BiasLayerTest/0.TestForwardEltwise -[ OK ] BiasLayerTest/0.TestForwardEltwise (0 ms) -[ RUN ] BiasLayerTest/0.TestForwardEltwiseInPlace -[ OK ] BiasLayerTest/0.TestForwardEltwiseInPlace (0 ms) -[ RUN ] BiasLayerTest/0.TestBackwardBroadcastMiddleInPlace -[ OK ] BiasLayerTest/0.TestBackwardBroadcastMiddleInPlace (0 ms) -[ RUN ] BiasLayerTest/0.TestGradientEltwiseWithParam -[ OK ] BiasLayerTest/0.TestGradientEltwiseWithParam (288 ms) -[ RUN ] BiasLayerTest/0.TestForwardBroadcastBegin -[ OK ] BiasLayerTest/0.TestForwardBroadcastBegin (0 ms) -[ RUN ] BiasLayerTest/0.TestForwardBroadcastMiddle -[ OK ] BiasLayerTest/0.TestForwardBroadcastMiddle (1 ms) -[ RUN ] BiasLayerTest/0.TestForwardBiasAxis2 -[ OK ] BiasLayerTest/0.TestForwardBiasAxis2 (0 ms) -[ RUN ] BiasLayerTest/0.TestGradientBiasAxis2 -[ OK ] BiasLayerTest/0.TestGradientBiasAxis2 (124 ms) -[ RUN ] BiasLayerTest/0.TestForwardBroadcastMiddleWithParam -[ OK ] BiasLayerTest/0.TestForwardBroadcastMiddleWithParam (0 ms) -[ RUN ] BiasLayerTest/0.TestGradientBias -[ OK ] BiasLayerTest/0.TestGradientBias (126 ms) -[ RUN ] BiasLayerTest/0.TestGradientBroadcastEnd -[ OK ] BiasLayerTest/0.TestGradientBroadcastEnd (187 ms) -[ RUN ] BiasLayerTest/0.TestGradientBroadcastMiddle -[ OK ] BiasLayerTest/0.TestGradientBroadcastMiddle (150 ms) -[ RUN ] BiasLayerTest/0.TestGradientEltwise -[ OK ] BiasLayerTest/0.TestGradientEltwise (6 ms) -[ RUN ] BiasLayerTest/0.TestGradientBroadcastBegin -[ OK ] BiasLayerTest/0.TestGradientBroadcastBegin (135 ms) -[ RUN ] BiasLayerTest/0.TestBackwardEltwiseInPlace -[ OK ] BiasLayerTest/0.TestBackwardEltwiseInPlace (0 ms) -[ RUN ] BiasLayerTest/0.TestForwardBroadcastMiddleInPlace -[ OK ] BiasLayerTest/0.TestForwardBroadcastMiddleInPlace (1 ms) -[ RUN ] BiasLayerTest/0.TestGradientBroadcastMiddleWithParam -[ OK ] BiasLayerTest/0.TestGradientBroadcastMiddleWithParam (148 ms) -[----------] 20 tests from BiasLayerTest/0 (1168 ms total) - -[----------] 5 tests from ImageDataLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] ImageDataLayerTest/0.TestShuffle -[ OK ] ImageDataLayerTest/0.TestShuffle (175 ms) -[ RUN ] ImageDataLayerTest/0.TestRead -[ OK ] ImageDataLayerTest/0.TestRead (181 ms) -[ RUN ] ImageDataLayerTest/0.TestResize -[ OK ] ImageDataLayerTest/0.TestResize (180 ms) -[ RUN ] ImageDataLayerTest/0.TestSpace -[ OK ] ImageDataLayerTest/0.TestSpace (50 ms) -[ RUN ] ImageDataLayerTest/0.TestReshape -[ OK ] ImageDataLayerTest/0.TestReshape (52 ms) -[----------] 5 tests from ImageDataLayerTest/0 (638 ms total) - -[----------] 2 tests from InternalThreadTest -[ RUN ] InternalThreadTest.TestStartAndExit -[ OK ] InternalThreadTest.TestStartAndExit (0 ms) -[ RUN ] InternalThreadTest.TestRandomSeed -[ OK ] InternalThreadTest.TestRandomSeed (1 ms) -[----------] 2 tests from InternalThreadTest (1 ms total) - -[----------] 9 tests from AccuracyLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] AccuracyLayerTest/1.TestSetupOutputPerClass -[ OK ] AccuracyLayerTest/1.TestSetupOutputPerClass (0 ms) -[ RUN ] AccuracyLayerTest/1.TestForwardPerClass -[ OK ] AccuracyLayerTest/1.TestForwardPerClass (1 ms) -[ RUN ] AccuracyLayerTest/1.TestSetup -[ OK ] AccuracyLayerTest/1.TestSetup (0 ms) -[ RUN ] AccuracyLayerTest/1.TestForwardTopK -[ OK ] AccuracyLayerTest/1.TestForwardTopK (6 ms) -[ RUN ] AccuracyLayerTest/1.TestForwardWithSpatialAxes -[ OK ] AccuracyLayerTest/1.TestForwardWithSpatialAxes (1 ms) -[ RUN ] AccuracyLayerTest/1.TestForwardPerClassWithIgnoreLabel -[ OK ] AccuracyLayerTest/1.TestForwardPerClassWithIgnoreLabel (1 ms) -[ RUN ] AccuracyLayerTest/1.TestSetupTopK -[ OK ] AccuracyLayerTest/1.TestSetupTopK (0 ms) -[ RUN ] AccuracyLayerTest/1.TestForwardIgnoreLabel -[ OK ] AccuracyLayerTest/1.TestForwardIgnoreLabel (0 ms) -[ RUN ] AccuracyLayerTest/1.TestForward -[ OK ] AccuracyLayerTest/1.TestForward (0 ms) -[----------] 9 tests from AccuracyLayerTest/1 (11 ms total) - -[----------] 5 tests from DeconvolutionLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] DeconvolutionLayerTest/1.TestSimpleDeconvolution -[ OK ] DeconvolutionLayerTest/1.TestSimpleDeconvolution (0 ms) -[ RUN ] DeconvolutionLayerTest/1.TestGradient3D -[ OK ] DeconvolutionLayerTest/1.TestGradient3D (428 ms) -[ RUN ] DeconvolutionLayerTest/1.TestGradient -[ OK ] DeconvolutionLayerTest/1.TestGradient (1091 ms) -[ RUN ] DeconvolutionLayerTest/1.TestSetup -[ OK ] DeconvolutionLayerTest/1.TestSetup (0 ms) -[ RUN ] DeconvolutionLayerTest/1.TestNDAgainst2D -[ OK ] DeconvolutionLayerTest/1.TestNDAgainst2D (1956 ms) -[----------] 5 tests from DeconvolutionLayerTest/1 (3475 ms total) - -[----------] 6 tests from MSRAFillerTest/0, where TypeParam = float -[ RUN ] MSRAFillerTest/0.TestFillFanIn -[ OK ] MSRAFillerTest/0.TestFillFanIn (97 ms) -[ RUN ] MSRAFillerTest/0.TestFill2D -[ OK ] MSRAFillerTest/0.TestFill2D (0 ms) -[ RUN ] MSRAFillerTest/0.TestFillAverage -[ OK ] MSRAFillerTest/0.TestFillAverage (97 ms) -[ RUN ] MSRAFillerTest/0.TestFill1D -[ OK ] MSRAFillerTest/0.TestFill1D (0 ms) -[ RUN ] MSRAFillerTest/0.TestFillFanOut -[ OK ] MSRAFillerTest/0.TestFillFanOut (97 ms) -[ RUN ] MSRAFillerTest/0.TestFill5D -[ OK ] MSRAFillerTest/0.TestFill5D (0 ms) -[----------] 6 tests from MSRAFillerTest/0 (291 ms total) - -[----------] 58 tests from NeuronLayerTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] NeuronLayerTest/1.TestExpLayerBase2 -[ OK ] NeuronLayerTest/1.TestExpLayerBase2 (0 ms) -[ RUN ] NeuronLayerTest/1.TestLogLayerBase2Shift1 -[ OK ] NeuronLayerTest/1.TestLogLayerBase2Shift1 (0 ms) -[ RUN ] NeuronLayerTest/1.TestAbsVal -[ OK ] NeuronLayerTest/1.TestAbsVal (0 ms) -[ RUN ] NeuronLayerTest/1.TestSigmoid -[ OK ] NeuronLayerTest/1.TestSigmoid (1 ms) -[ RUN ] NeuronLayerTest/1.TestLogGradientBase2 -[ OK ] NeuronLayerTest/1.TestLogGradientBase2 (5 ms) -[ RUN ] NeuronLayerTest/1.TestSwishAsLinearGradient -[ OK ] NeuronLayerTest/1.TestSwishAsLinearGradient (3 ms) -[ RUN ] NeuronLayerTest/1.TestLogLayerBase2 -[ OK ] NeuronLayerTest/1.TestLogLayerBase2 (0 ms) -[ RUN ] NeuronLayerTest/1.TestSwishWithBetaGradient -[ OK ] NeuronLayerTest/1.TestSwishWithBetaGradient (7 ms) -[ RUN ] NeuronLayerTest/1.TestPReLUForward -[ OK ] NeuronLayerTest/1.TestPReLUForward (0 ms) -[ RUN ] NeuronLayerTest/1.TestExpLayer -[ OK ] NeuronLayerTest/1.TestExpLayer (0 ms) -[ RUN ] NeuronLayerTest/1.TestSwish -[ OK ] NeuronLayerTest/1.TestSwish (0 ms) -[ RUN ] NeuronLayerTest/1.TestLogLayerBase2Shift1Scale3 -[ OK ] NeuronLayerTest/1.TestLogLayerBase2Shift1Scale3 (0 ms) -[ RUN ] NeuronLayerTest/1.TestExpLayerBase2Scale3 -[ OK ] NeuronLayerTest/1.TestExpLayerBase2Scale3 (0 ms) -[ RUN ] NeuronLayerTest/1.TestTanH -[ OK ] NeuronLayerTest/1.TestTanH (0 ms) -[ RUN ] NeuronLayerTest/1.TestReLUWithNegativeSlope -[ OK ] NeuronLayerTest/1.TestReLUWithNegativeSlope (0 ms) -[ RUN ] NeuronLayerTest/1.TestLogLayerBase2Scale3 -[ OK ] NeuronLayerTest/1.TestLogLayerBase2Scale3 (0 ms) -[ RUN ] NeuronLayerTest/1.TestExpGradientBase2Scale3 -[ OK ] NeuronLayerTest/1.TestExpGradientBase2Scale3 (5 ms) -[ RUN ] NeuronLayerTest/1.TestLogGradient -[ OK ] NeuronLayerTest/1.TestLogGradient (5 ms) -[ RUN ] NeuronLayerTest/1.TestExpLayerWithShift -[ OK ] NeuronLayerTest/1.TestExpLayerWithShift (0 ms) -[ RUN ] NeuronLayerTest/1.TestReLUGradient -[ OK ] NeuronLayerTest/1.TestReLUGradient (3 ms) -[ RUN ] NeuronLayerTest/1.TestReLUGradientWithNegativeSlope -[ OK ] NeuronLayerTest/1.TestReLUGradientWithNegativeSlope (3 ms) -[ RUN ] NeuronLayerTest/1.TestExpGradient -[ OK ] NeuronLayerTest/1.TestExpGradient (4 ms) -[ RUN ] NeuronLayerTest/1.TestDropoutThreeQuarters -[ OK ] NeuronLayerTest/1.TestDropoutThreeQuarters (0 ms) -[ RUN ] NeuronLayerTest/1.TestSwishGradient -[ OK ] NeuronLayerTest/1.TestSwishGradient (8 ms) -[ RUN ] NeuronLayerTest/1.TestExpLayerBase2Shift1Scale3 -[ OK ] NeuronLayerTest/1.TestExpLayerBase2Shift1Scale3 (0 ms) -[ RUN ] NeuronLayerTest/1.TestExpGradientBase2 -[ OK ] NeuronLayerTest/1.TestExpGradientBase2 (4 ms) -[ RUN ] NeuronLayerTest/1.TestPReLUGradient -[ OK ] NeuronLayerTest/1.TestPReLUGradient (155 ms) -[ RUN ] NeuronLayerTest/1.TestLogGradientBase2Shift1Scale3 -[ OK ] NeuronLayerTest/1.TestLogGradientBase2Shift1Scale3 (5 ms) -[ RUN ] NeuronLayerTest/1.TestClip -[ OK ] NeuronLayerTest/1.TestClip (1 ms) -[ RUN ] NeuronLayerTest/1.TestSigmoidGradient -[ OK ] NeuronLayerTest/1.TestSigmoidGradient (6 ms) -[ RUN ] NeuronLayerTest/1.TestReLU -[ OK ] NeuronLayerTest/1.TestReLU (1 ms) -[ RUN ] NeuronLayerTest/1.TestPReLUParam -[ OK ] NeuronLayerTest/1.TestPReLUParam (0 ms) -[ RUN ] NeuronLayerTest/1.TestLogLayer -[ OK ] NeuronLayerTest/1.TestLogLayer (0 ms) -[ RUN ] NeuronLayerTest/1.TestDropoutTestPhase -[ OK ] NeuronLayerTest/1.TestDropoutTestPhase (0 ms) -[ RUN ] NeuronLayerTest/1.TestELUasReLUGradient -[ OK ] NeuronLayerTest/1.TestELUasReLUGradient (4 ms) -[ RUN ] NeuronLayerTest/1.TestBNLLGradient -[ OK ] NeuronLayerTest/1.TestBNLLGradient (6 ms) -[ RUN ] NeuronLayerTest/1.TestExpLayerBase2Shift1 -[ OK ] NeuronLayerTest/1.TestExpLayerBase2Shift1 (0 ms) -[ RUN ] NeuronLayerTest/1.TestDropoutGradient -[ OK ] NeuronLayerTest/1.TestDropoutGradient (4 ms) -[ RUN ] NeuronLayerTest/1.TestClipGradient -[ OK ] NeuronLayerTest/1.TestClipGradient (3 ms) -[ RUN ] NeuronLayerTest/1.TestDropoutHalf -[ OK ] NeuronLayerTest/1.TestDropoutHalf (0 ms) -[ RUN ] NeuronLayerTest/1.TestSwishAsLinear -[ OK ] NeuronLayerTest/1.TestSwishAsLinear (0 ms) -[ RUN ] NeuronLayerTest/1.TestPReLUForwardChannelShared -[ OK ] NeuronLayerTest/1.TestPReLUForwardChannelShared (0 ms) -[ RUN ] NeuronLayerTest/1.TestPReLUInPlace -[ OK ] NeuronLayerTest/1.TestPReLUInPlace (1 ms) -[ RUN ] NeuronLayerTest/1.TestTanHGradient -[ OK ] NeuronLayerTest/1.TestTanHGradient (7 ms) -[ RUN ] NeuronLayerTest/1.TestLogGradientBase2Scale3 -[ OK ] NeuronLayerTest/1.TestLogGradientBase2Scale3 (6 ms) -[ RUN ] NeuronLayerTest/1.TestAbsGradient -[ OK ] NeuronLayerTest/1.TestAbsGradient (3 ms) -[ RUN ] NeuronLayerTest/1.TestLogGradientBase2Shift1 -[ OK ] NeuronLayerTest/1.TestLogGradientBase2Shift1 (5 ms) -[ RUN ] NeuronLayerTest/1.TestExpGradientWithShift -[ OK ] NeuronLayerTest/1.TestExpGradientWithShift (5 ms) -[ RUN ] NeuronLayerTest/1.TestELUGradient -[ OK ] NeuronLayerTest/1.TestELUGradient (3 ms) -[ RUN ] NeuronLayerTest/1.TestExpGradientBase2Shift1Scale3 -[ OK ] NeuronLayerTest/1.TestExpGradientBase2Shift1Scale3 (5 ms) -[ RUN ] NeuronLayerTest/1.TestExpGradientBase2Shift1 -[ OK ] NeuronLayerTest/1.TestExpGradientBase2Shift1 (4 ms) -[ RUN ] NeuronLayerTest/1.TestBNLL -[ OK ] NeuronLayerTest/1.TestBNLL (0 ms) -[ RUN ] NeuronLayerTest/1.TestDropoutGradientTest -[ OK ] NeuronLayerTest/1.TestDropoutGradientTest (3 ms) -[ RUN ] NeuronLayerTest/1.TestPReLUGradientChannelShared -[ OK ] NeuronLayerTest/1.TestPReLUGradientChannelShared (152 ms) -[ RUN ] NeuronLayerTest/1.TestELU -[ OK ] NeuronLayerTest/1.TestELU (0 ms) -[ RUN ] NeuronLayerTest/1.TestPReLUConsistencyReLU -[ OK ] NeuronLayerTest/1.TestPReLUConsistencyReLU (0 ms) -[ RUN ] NeuronLayerTest/1.TestELUasReLU -[ OK ] NeuronLayerTest/1.TestELUasReLU (0 ms) -[ RUN ] NeuronLayerTest/1.TestSwishWithBeta -[ OK ] NeuronLayerTest/1.TestSwishWithBeta (0 ms) -[----------] 58 tests from NeuronLayerTest/1 (430 ms total) - -[----------] 9 tests from AdaGradSolverTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] AdaGradSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare -[ OK ] AdaGradSolverTest/0.TestLeastSquaresUpdateWithEverythingAccumShare (7 ms) -[ RUN ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdate -[ OK ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdate (22 ms) -[ RUN ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdateWithEverything -[ OK ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdateWithEverything (109 ms) -[ RUN ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdateWithEverythingShare -[ OK ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdateWithEverythingShare (118 ms) -[ RUN ] AdaGradSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum -[ OK ] AdaGradSolverTest/0.TestLeastSquaresUpdateWithEverythingAccum (5 ms) -[ RUN ] AdaGradSolverTest/0.TestSnapshot -[ OK ] AdaGradSolverTest/0.TestSnapshot (32 ms) -[ RUN ] AdaGradSolverTest/0.TestSnapshotShare -[ OK ] AdaGradSolverTest/0.TestSnapshotShare (41 ms) -[ RUN ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdateWithWeightDecay -[ OK ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdateWithWeightDecay (22 ms) -[ RUN ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdateLROneHundredth -[ OK ] AdaGradSolverTest/0.TestAdaGradLeastSquaresUpdateLROneHundredth (21 ms) -[----------] 9 tests from AdaGradSolverTest/0 (377 ms total) - -[----------] 9 tests from InnerProductLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] InnerProductLayerTest/0.TestForwardNoBatch -[ OK ] InnerProductLayerTest/0.TestForwardNoBatch (0 ms) -[ RUN ] InnerProductLayerTest/0.TestForward -[ OK ] InnerProductLayerTest/0.TestForward (0 ms) -[ RUN ] InnerProductLayerTest/0.TestSetUpTransposeFalse -[ OK ] InnerProductLayerTest/0.TestSetUpTransposeFalse (0 ms) -[ RUN ] InnerProductLayerTest/0.TestGradientTranspose -[ OK ] InnerProductLayerTest/0.TestGradientTranspose (242 ms) -[ RUN ] InnerProductLayerTest/0.TestSetUpTransposeTrue -[ OK ] InnerProductLayerTest/0.TestSetUpTransposeTrue (0 ms) -[ RUN ] InnerProductLayerTest/0.TestGradient -[ OK ] InnerProductLayerTest/0.TestGradient (207 ms) -[ RUN ] InnerProductLayerTest/0.TestForwardTranspose -[ OK ] InnerProductLayerTest/0.TestForwardTranspose (1 ms) -[ RUN ] InnerProductLayerTest/0.TestSetUp -[ OK ] InnerProductLayerTest/0.TestSetUp (0 ms) -[ RUN ] InnerProductLayerTest/0.TestBackwardTranspose -[ OK ] InnerProductLayerTest/0.TestBackwardTranspose (0 ms) -[----------] 9 tests from InnerProductLayerTest/0 (451 ms total) - -[----------] 5 tests from DBTest/0, where TypeParam = caffe::TypeLevelDB -[ RUN ] DBTest/0.TestNext -[ OK ] DBTest/0.TestNext (24 ms) -[ RUN ] DBTest/0.TestWrite -[ OK ] DBTest/0.TestWrite (25 ms) -[ RUN ] DBTest/0.TestSeekToFirst -[ OK ] DBTest/0.TestSeekToFirst (26 ms) -[ RUN ] DBTest/0.TestGetDB -[ OK ] DBTest/0.TestGetDB (17 ms) -[ RUN ] DBTest/0.TestKeyValue -[ OK ] DBTest/0.TestKeyValue (25 ms) -[----------] 5 tests from DBTest/0 (118 ms total) +[----------] 14 tests from DataLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] DataLayerTest/1.TestReshapeLevelDB +[ OK ] DataLayerTest/1.TestReshapeLevelDB (3 ms) +[ RUN ] DataLayerTest/1.TestReadCropTrainSequenceSeededLevelDB +[ OK ] DataLayerTest/1.TestReadCropTrainSequenceSeededLevelDB (3 ms) +[ RUN ] DataLayerTest/1.TestSkipLMDB +[ OK ] DataLayerTest/1.TestSkipLMDB (9 ms) +[ RUN ] DataLayerTest/1.TestReadCropTestLMDB +[ OK ] DataLayerTest/1.TestReadCropTestLMDB (1 ms) +[ RUN ] DataLayerTest/1.TestReadCropTrainLevelDB +[ OK ] DataLayerTest/1.TestReadCropTrainLevelDB (3 ms) +[ RUN ] DataLayerTest/1.TestReadLMDB +[ OK ] DataLayerTest/1.TestReadLMDB (8 ms) +[ RUN ] DataLayerTest/1.TestReadCropTestLevelDB +[ OK ] DataLayerTest/1.TestReadCropTestLevelDB (2 ms) +[ RUN ] DataLayerTest/1.TestReadCropTrainLMDB +[ OK ] DataLayerTest/1.TestReadCropTrainLMDB (7 ms) +[ RUN ] DataLayerTest/1.TestReadLevelDB +[ OK ] DataLayerTest/1.TestReadLevelDB (12 ms) +[ RUN ] DataLayerTest/1.TestReadCropTrainSequenceSeededLMDB +[ OK ] DataLayerTest/1.TestReadCropTrainSequenceSeededLMDB (2 ms) +[ RUN ] DataLayerTest/1.TestReshapeLMDB +[ OK ] DataLayerTest/1.TestReshapeLMDB (1 ms) +[ RUN ] DataLayerTest/1.TestReadCropTrainSequenceUnseededLMDB +[ OK ] DataLayerTest/1.TestReadCropTrainSequenceUnseededLMDB (1 ms) +[ RUN ] DataLayerTest/1.TestSkipLevelDB +[ OK ] DataLayerTest/1.TestSkipLevelDB (21 ms) +[ RUN ] DataLayerTest/1.TestReadCropTrainSequenceUnseededLevelDB +[ OK ] DataLayerTest/1.TestReadCropTrainSequenceUnseededLevelDB (4 ms) +[----------] 14 tests from DataLayerTest/1 (77 ms total) -[----------] 1 test from HDF5OutputLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] HDF5OutputLayerTest/0.TestForward -[ OK ] HDF5OutputLayerTest/0.TestForward (5 ms) -[----------] 1 test from HDF5OutputLayerTest/0 (5 ms total) +[----------] 10 tests from ConcatLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] ConcatLayerTest/1.TestGradientChannels +[ OK ] ConcatLayerTest/1.TestGradientChannels (6 ms) +[ RUN ] ConcatLayerTest/1.TestForwardTrivial +[ OK ] ConcatLayerTest/1.TestForwardTrivial (0 ms) +[ RUN ] ConcatLayerTest/1.TestSetupNum +[ OK ] ConcatLayerTest/1.TestSetupNum (0 ms) +[ RUN ] ConcatLayerTest/1.TestGradientChannelsBottomOneOnly +[ OK ] ConcatLayerTest/1.TestGradientChannelsBottomOneOnly (4 ms) +[ RUN ] ConcatLayerTest/1.TestForwardNum +[ OK ] ConcatLayerTest/1.TestForwardNum (0 ms) +[ RUN ] ConcatLayerTest/1.TestGradientNum +[ OK ] ConcatLayerTest/1.TestGradientNum (12 ms) +[ RUN ] ConcatLayerTest/1.TestGradientTrivial +[ OK ] ConcatLayerTest/1.TestGradientTrivial (5 ms) +[ RUN ] ConcatLayerTest/1.TestSetupChannels +[ OK ] ConcatLayerTest/1.TestSetupChannels (0 ms) +[ RUN ] ConcatLayerTest/1.TestSetupChannelsNegativeIndexing +[ OK ] ConcatLayerTest/1.TestSetupChannelsNegativeIndexing (0 ms) +[ RUN ] ConcatLayerTest/1.TestForwardChannels +[ OK ] ConcatLayerTest/1.TestForwardChannels (1 ms) +[----------] 10 tests from ConcatLayerTest/1 (29 ms total) -[----------] 2 tests from HDF5DataLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] HDF5DataLayerTest/0.TestRead -[ OK ] HDF5DataLayerTest/0.TestRead (13 ms) -[ RUN ] HDF5DataLayerTest/0.TestSkip -[ OK ] HDF5DataLayerTest/0.TestSkip (43 ms) -[----------] 2 tests from HDF5DataLayerTest/0 (56 ms total) +[----------] 12 tests from SGDSolverTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum +[ OK ] SGDSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum (6 ms) +[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdateWithMomentum +[ OK ] SGDSolverTest/1.TestLeastSquaresUpdateWithMomentum (51 ms) +[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdate +[ OK ] SGDSolverTest/1.TestLeastSquaresUpdate (23 ms) +[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdateWithEverythingShare +[ OK ] SGDSolverTest/1.TestLeastSquaresUpdateWithEverythingShare (123 ms) +[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdateWithEverything +[ OK ] SGDSolverTest/1.TestLeastSquaresUpdateWithEverything (116 ms) +[ RUN ] SGDSolverTest/1.TestSnapshotShare +[ OK ] SGDSolverTest/1.TestSnapshotShare (50 ms) +[ RUN ] SGDSolverTest/1.TestSnapshot +[ OK ] SGDSolverTest/1.TestSnapshot (40 ms) +[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdateLROneHundredth +[ OK ] SGDSolverTest/1.TestLeastSquaresUpdateLROneHundredth (23 ms) +[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdateWithWeightDecay +[ OK ] SGDSolverTest/1.TestLeastSquaresUpdateWithWeightDecay (46 ms) +[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdateWithWeightDecayMultiIter +[ OK ] SGDSolverTest/1.TestLeastSquaresUpdateWithWeightDecayMultiIter (116 ms) +[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare +[ OK ] SGDSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare (8 ms) +[ RUN ] SGDSolverTest/1.TestLeastSquaresUpdateWithMomentumMultiIter +[ OK ] SGDSolverTest/1.TestLeastSquaresUpdateWithMomentumMultiIter (116 ms) +[----------] 12 tests from SGDSolverTest/1 (718 ms total) -[----------] 8 tests from RMSPropSolverTest/1, where TypeParam = caffe::CPUDevice -[ RUN ] RMSPropSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare -[ OK ] RMSPropSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare (8 ms) -[ RUN ] RMSPropSolverTest/1.TestRMSPropLeastSquaresUpdateWithEverythingShare -[ OK ] RMSPropSolverTest/1.TestRMSPropLeastSquaresUpdateWithEverythingShare (124 ms) -[ RUN ] RMSPropSolverTest/1.TestRMSPropLeastSquaresUpdateWithRmsDecay -[ OK ] RMSPropSolverTest/1.TestRMSPropLeastSquaresUpdateWithRmsDecay (113 ms) -[ RUN ] RMSPropSolverTest/1.TestRMSPropLeastSquaresUpdateWithWeightDecay -[ OK ] RMSPropSolverTest/1.TestRMSPropLeastSquaresUpdateWithWeightDecay (22 ms) -[ RUN ] RMSPropSolverTest/1.TestSnapshotShare -[ OK ] RMSPropSolverTest/1.TestSnapshotShare (48 ms) -[ RUN ] RMSPropSolverTest/1.TestRMSPropLeastSquaresUpdateWithEverything -[ OK ] RMSPropSolverTest/1.TestRMSPropLeastSquaresUpdateWithEverything (114 ms) -[ RUN ] RMSPropSolverTest/1.TestSnapshot -[ OK ] RMSPropSolverTest/1.TestSnapshot (36 ms) -[ RUN ] RMSPropSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum -[ OK ] RMSPropSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum (6 ms) -[----------] 8 tests from RMSPropSolverTest/1 (473 ms total) +[----------] 3 tests from FilterLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] FilterLayerTest/1.TestReshape +[ OK ] FilterLayerTest/1.TestReshape (0 ms) +[ RUN ] FilterLayerTest/1.TestForward +[ OK ] FilterLayerTest/1.TestForward (0 ms) +[ RUN ] FilterLayerTest/1.TestGradient +[ OK ] FilterLayerTest/1.TestGradient (456 ms) +[----------] 3 tests from FilterLayerTest/1 (457 ms total) -[----------] 12 tests from ReshapeLayerTest/0, where TypeParam = caffe::CPUDevice -[ RUN ] ReshapeLayerTest/0.TestInsertSingletonAxesEnd -[ OK ] ReshapeLayerTest/0.TestInsertSingletonAxesEnd (0 ms) -[ RUN ] ReshapeLayerTest/0.TestForward -[ OK ] ReshapeLayerTest/0.TestForward (0 ms) -[ RUN ] ReshapeLayerTest/0.TestInsertSingletonAxesStart -[ OK ] ReshapeLayerTest/0.TestInsertSingletonAxesStart (0 ms) -[ RUN ] ReshapeLayerTest/0.TestGradient -[ OK ] ReshapeLayerTest/0.TestGradient (4 ms) -[ RUN ] ReshapeLayerTest/0.TestInferenceOfUnspecified -[ OK ] ReshapeLayerTest/0.TestInferenceOfUnspecified (0 ms) -[ RUN ] ReshapeLayerTest/0.TestForwardAfterReshape -[ OK ] ReshapeLayerTest/0.TestForwardAfterReshape (0 ms) -[ RUN ] ReshapeLayerTest/0.TestFlattenValues -[ OK ] ReshapeLayerTest/0.TestFlattenValues (1 ms) -[ RUN ] ReshapeLayerTest/0.TestInsertSingletonAxesMiddle -[ OK ] ReshapeLayerTest/0.TestInsertSingletonAxesMiddle (0 ms) -[ RUN ] ReshapeLayerTest/0.TestInferenceOfUnspecifiedWithStartAxis -[ OK ] ReshapeLayerTest/0.TestInferenceOfUnspecifiedWithStartAxis (0 ms) -[ RUN ] ReshapeLayerTest/0.TestFlattenOutputSizes -[ OK ] ReshapeLayerTest/0.TestFlattenOutputSizes (0 ms) -[ RUN ] ReshapeLayerTest/0.TestFlattenMiddle -[ OK ] ReshapeLayerTest/0.TestFlattenMiddle (0 ms) -[ RUN ] ReshapeLayerTest/0.TestCopyDimensions -[ OK ] ReshapeLayerTest/0.TestCopyDimensions (0 ms) -[----------] 12 tests from ReshapeLayerTest/0 (5 ms total) +[----------] 9 tests from AdaGradSolverTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] AdaGradSolverTest/1.TestSnapshotShare +[ OK ] AdaGradSolverTest/1.TestSnapshotShare (51 ms) +[ RUN ] AdaGradSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare +[ OK ] AdaGradSolverTest/1.TestLeastSquaresUpdateWithEverythingAccumShare (9 ms) +[ RUN ] AdaGradSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum +[ OK ] AdaGradSolverTest/1.TestLeastSquaresUpdateWithEverythingAccum (7 ms) +[ RUN ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdateWithEverythingShare +[ OK ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdateWithEverythingShare (125 ms) +[ RUN ] AdaGradSolverTest/1.TestSnapshot +[ OK ] AdaGradSolverTest/1.TestSnapshot (42 ms) +[ RUN ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdateWithEverything +[ OK ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdateWithEverything (117 ms) +[ RUN ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdateWithWeightDecay +[ OK ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdateWithWeightDecay (23 ms) +[ RUN ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdateLROneHundredth +[ OK ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdateLROneHundredth (24 ms) +[ RUN ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdate +[ OK ] AdaGradSolverTest/1.TestAdaGradLeastSquaresUpdate (23 ms) +[----------] 9 tests from AdaGradSolverTest/1 (421 ms total) -[----------] 4 tests from NetUpgradeTest -[ RUN ] NetUpgradeTest.TestImageNet -[ OK ] NetUpgradeTest.TestImageNet (5 ms) -[ RUN ] NetUpgradeTest.TestUpgradeV1LayerType -[ OK ] NetUpgradeTest.TestUpgradeV1LayerType (3 ms) -[ RUN ] NetUpgradeTest.TestSimple -[ OK ] NetUpgradeTest.TestSimple (1 ms) -[ RUN ] NetUpgradeTest.TestAllParams -[ OK ] NetUpgradeTest.TestAllParams (2 ms) -[----------] 4 tests from NetUpgradeTest (11 ms total) +[----------] 27 tests from ReductionLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] ReductionLayerTest/1.TestSumOfSquaresCoeffAxis1 +[ OK ] ReductionLayerTest/1.TestSumOfSquaresCoeffAxis1 (0 ms) +[ RUN ] ReductionLayerTest/1.TestSumCoeffAxis1Gradient +[ OK ] ReductionLayerTest/1.TestSumCoeffAxis1Gradient (2 ms) +[ RUN ] ReductionLayerTest/1.TestSetUpWithAxis1 +[ OK ] ReductionLayerTest/1.TestSetUpWithAxis1 (0 ms) +[ RUN ] ReductionLayerTest/1.TestSumOfSquaresCoeffAxis1Gradient +[ OK ] ReductionLayerTest/1.TestSumOfSquaresCoeffAxis1Gradient (3 ms) +[ RUN ] ReductionLayerTest/1.TestSetUp +[ OK ] ReductionLayerTest/1.TestSetUp (0 ms) +[ RUN ] ReductionLayerTest/1.TestSumCoeffGradient +[ OK ] ReductionLayerTest/1.TestSumCoeffGradient (1 ms) +[ RUN ] ReductionLayerTest/1.TestSumOfSquaresCoeffGradient +[ OK ] ReductionLayerTest/1.TestSumOfSquaresCoeffGradient (1 ms) +[ RUN ] ReductionLayerTest/1.TestMeanGradient +[ OK ] ReductionLayerTest/1.TestMeanGradient (1 ms) +[ RUN ] ReductionLayerTest/1.TestAbsSumCoeffAxis1Gradient +[ OK ] ReductionLayerTest/1.TestAbsSumCoeffAxis1Gradient (3 ms) +[ RUN ] ReductionLayerTest/1.TestAbsSumCoeff +[ OK ] ReductionLayerTest/1.TestAbsSumCoeff (0 ms) +[ RUN ] ReductionLayerTest/1.TestMeanCoeffAxis1 +[ OK ] ReductionLayerTest/1.TestMeanCoeffAxis1 (0 ms) +[ RUN ] ReductionLayerTest/1.TestAbsSum +[ OK ] ReductionLayerTest/1.TestAbsSum (0 ms) +[ RUN ] ReductionLayerTest/1.TestMean +[ OK ] ReductionLayerTest/1.TestMean (0 ms) +[ RUN ] ReductionLayerTest/1.TestAbsSumCoeffAxis1 +[ OK ] ReductionLayerTest/1.TestAbsSumCoeffAxis1 (0 ms) +[ RUN ] ReductionLayerTest/1.TestSumCoeffAxis1 +[ OK ] ReductionLayerTest/1.TestSumCoeffAxis1 (0 ms) +[ RUN ] ReductionLayerTest/1.TestAbsSumGradient +[ OK ] ReductionLayerTest/1.TestAbsSumGradient (1 ms) +[ RUN ] ReductionLayerTest/1.TestMeanCoeffGradient +[ OK ] ReductionLayerTest/1.TestMeanCoeffGradient (1 ms) +[ RUN ] ReductionLayerTest/1.TestMeanCoeff +[ OK ] ReductionLayerTest/1.TestMeanCoeff (0 ms) +[ RUN ] ReductionLayerTest/1.TestSumGradient +[ OK ] ReductionLayerTest/1.TestSumGradient (2 ms) +[ RUN ] ReductionLayerTest/1.TestSumOfSquares +[ OK ] ReductionLayerTest/1.TestSumOfSquares (0 ms) +[ RUN ] ReductionLayerTest/1.TestMeanCoeffGradientAxis1 +[ OK ] ReductionLayerTest/1.TestMeanCoeffGradientAxis1 (2 ms) +[ RUN ] ReductionLayerTest/1.TestSetUpWithAxis2 +[ OK ] ReductionLayerTest/1.TestSetUpWithAxis2 (0 ms) +[ RUN ] ReductionLayerTest/1.TestSum +[ OK ] ReductionLayerTest/1.TestSum (0 ms) +[ RUN ] ReductionLayerTest/1.TestSumOfSquaresCoeff +[ OK ] ReductionLayerTest/1.TestSumOfSquaresCoeff (0 ms) +[ RUN ] ReductionLayerTest/1.TestSumOfSquaresGradient +[ OK ] ReductionLayerTest/1.TestSumOfSquaresGradient (1 ms) +[ RUN ] ReductionLayerTest/1.TestSumCoeff +[ OK ] ReductionLayerTest/1.TestSumCoeff (0 ms) +[ RUN ] ReductionLayerTest/1.TestAbsSumCoeffGradient +[ OK ] ReductionLayerTest/1.TestAbsSumCoeffGradient (1 ms) +[----------] 27 tests from ReductionLayerTest/1 (19 ms total) -[----------] 1 test from CPUStochasticPoolingLayerTest/1, where TypeParam = double -[ RUN ] CPUStochasticPoolingLayerTest/1.TestSetup -[ OK ] CPUStochasticPoolingLayerTest/1.TestSetup (1 ms) -[----------] 1 test from CPUStochasticPoolingLayerTest/1 (1 ms total) +[----------] 2 tests from EuclideanLossLayerTest/1, where TypeParam = caffe::CPUDevice +[ RUN ] EuclideanLossLayerTest/1.TestForward +[ OK ] EuclideanLossLayerTest/1.TestForward (1 ms) +[ RUN ] EuclideanLossLayerTest/1.TestGradient +[ OK ] EuclideanLossLayerTest/1.TestGradient (1 ms) +[----------] 2 tests from EuclideanLossLayerTest/1 (2 ms total) [----------] Global test environment tear-down -[==========] 1162 tests from 152 test cases ran. (103207 ms total) +[==========] 1162 tests from 152 test cases ran. (107131 ms total) [ PASSED ] 1162 tests. make[5]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' [100%] Built target runtest @@ -14816,11 +14883,11 @@ .../build/caffe-1.0.0+git20180821.99bd997/python/caffe/test/test_coord_map.py:45: DeprecationWarning: Please use assertEqual instead. self.assertEquals(ax, 1) ................WARNING: Logging before InitGoogleLogging() is written to STDERR -I0819 13:44:51.978968 13551 net.cpp:296] The NetState phase (1) differed from the phase (0) specified by a rule in layer train_data -I0819 13:44:51.979040 13551 net.cpp:332] The NetState did not contain stage 'val' specified by a rule in layer val_data -I0819 13:44:51.979063 13551 net.cpp:296] The NetState phase (1) differed from the phase (0) specified by a rule in layer loss -I0819 13:44:51.979074 13551 net.cpp:332] The NetState did not contain stage 'val' specified by a rule in layer loss -I0819 13:44:51.979087 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.374714 31304 net.cpp:296] The NetState phase (1) differed from the phase (0) specified by a rule in layer train_data +I0718 09:46:46.374802 31304 net.cpp:332] The NetState did not contain stage 'val' specified by a rule in layer val_data +I0718 09:46:46.374825 31304 net.cpp:296] The NetState phase (1) differed from the phase (0) specified by a rule in layer loss +I0718 09:46:46.374836 31304 net.cpp:332] The NetState did not contain stage 'val' specified by a rule in layer loss +I0718 09:46:46.374850 31304 net.cpp:53] Initializing net from parameters: state { phase: TEST level: 0 @@ -14862,35 +14929,35 @@ stage: "deploy" } } -I0819 13:44:51.979269 13551 layer_factory.hpp:77] Creating layer deploy_data -I0819 13:44:51.979317 13551 net.cpp:86] Creating Layer deploy_data -I0819 13:44:51.979336 13551 net.cpp:382] deploy_data -> data -I0819 13:44:51.979413 13551 net.cpp:124] Setting up deploy_data -I0819 13:44:51.979427 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:51.979444 13551 net.cpp:139] Memory required for data: 400 -I0819 13:44:51.979455 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:51.979473 13551 net.cpp:86] Creating Layer ip -I0819 13:44:51.979485 13551 net.cpp:408] ip <- data -I0819 13:44:51.979501 13551 net.cpp:382] ip -> ip -I0819 13:44:51.979557 13551 net.cpp:124] Setting up ip -I0819 13:44:51.979569 13551 net.cpp:131] Top shape: 1 2 (2) -I0819 13:44:51.979580 13551 net.cpp:139] Memory required for data: 408 -I0819 13:44:51.979606 13551 layer_factory.hpp:77] Creating layer pred -I0819 13:44:51.979626 13551 net.cpp:86] Creating Layer pred -I0819 13:44:51.979637 13551 net.cpp:408] pred <- ip -I0819 13:44:51.979652 13551 net.cpp:382] pred -> pred -I0819 13:44:51.979691 13551 net.cpp:124] Setting up pred -I0819 13:44:51.979703 13551 net.cpp:131] Top shape: 1 2 (2) -I0819 13:44:51.979714 13551 net.cpp:139] Memory required for data: 416 -I0819 13:44:51.979723 13551 net.cpp:202] pred does not need backward computation. -I0819 13:44:51.979734 13551 net.cpp:202] ip does not need backward computation. -I0819 13:44:51.979743 13551 net.cpp:202] deploy_data does not need backward computation. -I0819 13:44:51.979753 13551 net.cpp:244] This network produces output pred -I0819 13:44:51.979766 13551 net.cpp:257] Network initialization done. -.I0819 13:44:51.982306 13551 net.cpp:296] The NetState phase (0) differed from the phase (1) specified by a rule in layer val_data -I0819 13:44:51.982345 13551 net.cpp:296] The NetState phase (0) differed from the phase (1) specified by a rule in layer deploy_data -I0819 13:44:51.982364 13551 net.cpp:296] The NetState phase (0) differed from the phase (1) specified by a rule in layer pred -I0819 13:44:51.982376 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.375036 31304 layer_factory.hpp:77] Creating layer deploy_data +I0718 09:46:46.375085 31304 net.cpp:86] Creating Layer deploy_data +I0718 09:46:46.375104 31304 net.cpp:382] deploy_data -> data +I0718 09:46:46.375185 31304 net.cpp:124] Setting up deploy_data +I0718 09:46:46.375349 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.375367 31304 net.cpp:139] Memory required for data: 400 +I0718 09:46:46.375377 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.375401 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.375414 31304 net.cpp:408] ip <- data +I0718 09:46:46.375432 31304 net.cpp:382] ip -> ip +I0718 09:46:46.375499 31304 net.cpp:124] Setting up ip +I0718 09:46:46.375515 31304 net.cpp:131] Top shape: 1 2 (2) +I0718 09:46:46.375530 31304 net.cpp:139] Memory required for data: 408 +I0718 09:46:46.375556 31304 layer_factory.hpp:77] Creating layer pred +I0718 09:46:46.375579 31304 net.cpp:86] Creating Layer pred +I0718 09:46:46.375591 31304 net.cpp:408] pred <- ip +I0718 09:46:46.375605 31304 net.cpp:382] pred -> pred +I0718 09:46:46.375646 31304 net.cpp:124] Setting up pred +I0718 09:46:46.375659 31304 net.cpp:131] Top shape: 1 2 (2) +I0718 09:46:46.375671 31304 net.cpp:139] Memory required for data: 416 +I0718 09:46:46.375680 31304 net.cpp:202] pred does not need backward computation. +I0718 09:46:46.375691 31304 net.cpp:202] ip does not need backward computation. +I0718 09:46:46.375701 31304 net.cpp:202] deploy_data does not need backward computation. +I0718 09:46:46.375710 31304 net.cpp:244] This network produces output pred +I0718 09:46:46.375725 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.378221 31304 net.cpp:296] The NetState phase (0) differed from the phase (1) specified by a rule in layer val_data +I0718 09:46:46.378249 31304 net.cpp:296] The NetState phase (0) differed from the phase (1) specified by a rule in layer deploy_data +I0718 09:46:46.378268 31304 net.cpp:296] The NetState phase (0) differed from the phase (1) specified by a rule in layer pred +I0718 09:46:46.378279 31304 net.cpp:53] Initializing net from parameters: state { phase: TRAIN level: 0 @@ -14944,41 +15011,41 @@ stage: "val" } } -I0819 13:44:51.982550 13551 layer_factory.hpp:77] Creating layer train_data -I0819 13:44:51.982586 13551 net.cpp:86] Creating Layer train_data -I0819 13:44:51.982604 13551 net.cpp:382] train_data -> data -I0819 13:44:51.982630 13551 net.cpp:382] train_data -> label -I0819 13:44:51.982666 13551 net.cpp:124] Setting up train_data -I0819 13:44:51.982676 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:51.982689 13551 net.cpp:131] Top shape: 1 1 1 1 (1) -I0819 13:44:51.982699 13551 net.cpp:139] Memory required for data: 404 -I0819 13:44:51.982708 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:51.982720 13551 net.cpp:86] Creating Layer ip -I0819 13:44:51.982729 13551 net.cpp:408] ip <- data -I0819 13:44:51.982743 13551 net.cpp:382] ip -> ip -I0819 13:44:51.982775 13551 net.cpp:124] Setting up ip -I0819 13:44:51.982785 13551 net.cpp:131] Top shape: 1 2 (2) -I0819 13:44:51.982796 13551 net.cpp:139] Memory required for data: 412 -I0819 13:44:51.982816 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:51.982836 13551 net.cpp:86] Creating Layer loss -I0819 13:44:51.982844 13551 net.cpp:408] loss <- ip -I0819 13:44:51.982854 13551 net.cpp:408] loss <- label -I0819 13:44:51.982869 13551 net.cpp:382] loss -> loss -I0819 13:44:51.982908 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:51.982949 13551 net.cpp:124] Setting up loss -I0819 13:44:51.982959 13551 net.cpp:131] Top shape: (1) -I0819 13:44:51.982970 13551 net.cpp:134] with loss weight 1 -I0819 13:44:51.983000 13551 net.cpp:139] Memory required for data: 416 -I0819 13:44:51.983011 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:51.983021 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:51.983031 13551 net.cpp:202] train_data does not need backward computation. -I0819 13:44:51.983039 13551 net.cpp:244] This network produces output loss -I0819 13:44:51.983053 13551 net.cpp:257] Network initialization done. -.I0819 13:44:51.990955 13551 net.cpp:296] The NetState phase (1) differed from the phase (0) specified by a rule in layer train_data -I0819 13:44:51.991009 13551 net.cpp:332] The NetState did not contain stage 'deploy' specified by a rule in layer deploy_data -I0819 13:44:51.991022 13551 net.cpp:296] The NetState phase (1) differed from the phase (0) specified by a rule in layer loss -I0819 13:44:51.991035 13551 net.cpp:332] The NetState did not contain stage 'deploy' specified by a rule in layer pred -I0819 13:44:51.991046 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.378456 31304 layer_factory.hpp:77] Creating layer train_data +I0718 09:46:46.378486 31304 net.cpp:86] Creating Layer train_data +I0718 09:46:46.378505 31304 net.cpp:382] train_data -> data +I0718 09:46:46.378536 31304 net.cpp:382] train_data -> label +I0718 09:46:46.378579 31304 net.cpp:124] Setting up train_data +I0718 09:46:46.378597 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.378614 31304 net.cpp:131] Top shape: 1 1 1 1 (1) +I0718 09:46:46.378626 31304 net.cpp:139] Memory required for data: 404 +I0718 09:46:46.378638 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.378656 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.378669 31304 net.cpp:408] ip <- data +I0718 09:46:46.378684 31304 net.cpp:382] ip -> ip +I0718 09:46:46.378726 31304 net.cpp:124] Setting up ip +I0718 09:46:46.378741 31304 net.cpp:131] Top shape: 1 2 (2) +I0718 09:46:46.378752 31304 net.cpp:139] Memory required for data: 412 +I0718 09:46:46.378777 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.378801 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.378814 31304 net.cpp:408] loss <- ip +I0718 09:46:46.378827 31304 net.cpp:408] loss <- label +I0718 09:46:46.378845 31304 net.cpp:382] loss -> loss +I0718 09:46:46.378893 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.378950 31304 net.cpp:124] Setting up loss +I0718 09:46:46.378966 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.378979 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.379014 31304 net.cpp:139] Memory required for data: 416 +I0718 09:46:46.379029 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.379041 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.379052 31304 net.cpp:202] train_data does not need backward computation. +I0718 09:46:46.379060 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.379076 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.380898 31304 net.cpp:296] The NetState phase (1) differed from the phase (0) specified by a rule in layer train_data +I0718 09:46:46.380936 31304 net.cpp:332] The NetState did not contain stage 'deploy' specified by a rule in layer deploy_data +I0718 09:46:46.380949 31304 net.cpp:296] The NetState phase (1) differed from the phase (0) specified by a rule in layer loss +I0718 09:46:46.380964 31304 net.cpp:332] The NetState did not contain stage 'deploy' specified by a rule in layer pred +I0718 09:46:46.380973 31304 net.cpp:53] Initializing net from parameters: state { phase: TEST level: 0 @@ -15032,39 +15099,39 @@ stage: "val" } } -I0819 13:44:51.991231 13551 layer_factory.hpp:77] Creating layer val_data -I0819 13:44:51.991263 13551 net.cpp:86] Creating Layer val_data -I0819 13:44:51.991284 13551 net.cpp:382] val_data -> data -I0819 13:44:51.991322 13551 net.cpp:382] val_data -> label -I0819 13:44:51.991364 13551 net.cpp:124] Setting up val_data -I0819 13:44:51.991377 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:51.991394 13551 net.cpp:131] Top shape: 1 1 1 1 (1) -I0819 13:44:51.991405 13551 net.cpp:139] Memory required for data: 404 -I0819 13:44:51.991415 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:51.991431 13551 net.cpp:86] Creating Layer ip -I0819 13:44:51.991443 13551 net.cpp:408] ip <- data -I0819 13:44:51.991459 13551 net.cpp:382] ip -> ip -I0819 13:44:51.991501 13551 net.cpp:124] Setting up ip -I0819 13:44:51.991514 13551 net.cpp:131] Top shape: 1 2 (2) -I0819 13:44:51.991524 13551 net.cpp:139] Memory required for data: 412 -I0819 13:44:51.991550 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:51.991575 13551 net.cpp:86] Creating Layer loss -I0819 13:44:51.991585 13551 net.cpp:408] loss <- ip -I0819 13:44:51.991597 13551 net.cpp:408] loss <- label -I0819 13:44:51.991613 13551 net.cpp:382] loss -> loss -I0819 13:44:51.991641 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:51.991680 13551 net.cpp:124] Setting up loss -I0819 13:44:51.991690 13551 net.cpp:131] Top shape: (1) -I0819 13:44:51.991701 13551 net.cpp:134] with loss weight 1 -I0819 13:44:51.991729 13551 net.cpp:139] Memory required for data: 416 -I0819 13:44:51.991737 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:51.991748 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:51.991757 13551 net.cpp:202] val_data does not need backward computation. -I0819 13:44:51.991765 13551 net.cpp:244] This network produces output loss -I0819 13:44:51.991780 13551 net.cpp:257] Network initialization done. -.I0819 13:44:51.993872 13551 net.cpp:306] The NetState level (0) is above the min_level (1) specified by a rule in layer Level1Only -I0819 13:44:51.993913 13551 net.cpp:306] The NetState level (0) is above the min_level (1) specified by a rule in layer Level>=1 -I0819 13:44:51.993924 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.381161 31304 layer_factory.hpp:77] Creating layer val_data +I0718 09:46:46.381191 31304 net.cpp:86] Creating Layer val_data +I0718 09:46:46.381211 31304 net.cpp:382] val_data -> data +I0718 09:46:46.381242 31304 net.cpp:382] val_data -> label +I0718 09:46:46.381280 31304 net.cpp:124] Setting up val_data +I0718 09:46:46.381294 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.381309 31304 net.cpp:131] Top shape: 1 1 1 1 (1) +I0718 09:46:46.381319 31304 net.cpp:139] Memory required for data: 404 +I0718 09:46:46.381328 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.381343 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.381356 31304 net.cpp:408] ip <- data +I0718 09:46:46.381371 31304 net.cpp:382] ip -> ip +I0718 09:46:46.381408 31304 net.cpp:124] Setting up ip +I0718 09:46:46.381420 31304 net.cpp:131] Top shape: 1 2 (2) +I0718 09:46:46.381430 31304 net.cpp:139] Memory required for data: 412 +I0718 09:46:46.381453 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.381475 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.381484 31304 net.cpp:408] loss <- ip +I0718 09:46:46.381496 31304 net.cpp:408] loss <- label +I0718 09:46:46.381511 31304 net.cpp:382] loss -> loss +I0718 09:46:46.381534 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.381574 31304 net.cpp:124] Setting up loss +I0718 09:46:46.381587 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.381598 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.381624 31304 net.cpp:139] Memory required for data: 416 +I0718 09:46:46.381635 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.381647 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.381657 31304 net.cpp:202] val_data does not need backward computation. +I0718 09:46:46.381666 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.381682 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.383440 31304 net.cpp:306] The NetState level (0) is above the min_level (1) specified by a rule in layer Level1Only +I0718 09:46:46.383477 31304 net.cpp:306] The NetState level (0) is above the min_level (1) specified by a rule in layer Level>=1 +I0718 09:46:46.383488 31304 net.cpp:53] Initializing net from parameters: state { phase: TEST level: 0 @@ -15116,55 +15183,55 @@ num_output: 1 } } -I0819 13:44:51.994133 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:51.994161 13551 net.cpp:86] Creating Layer data -I0819 13:44:51.994177 13551 net.cpp:382] data -> data -I0819 13:44:51.994215 13551 net.cpp:124] Setting up data -I0819 13:44:51.994227 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:51.994240 13551 net.cpp:139] Memory required for data: 400 -I0819 13:44:51.994248 13551 layer_factory.hpp:77] Creating layer data_data_0_split -I0819 13:44:51.994267 13551 net.cpp:86] Creating Layer data_data_0_split -I0819 13:44:51.994277 13551 net.cpp:408] data_data_0_split <- data -I0819 13:44:51.994292 13551 net.cpp:382] data_data_0_split -> data_data_0_split_0 -I0819 13:44:51.994307 13551 net.cpp:382] data_data_0_split -> data_data_0_split_1 -I0819 13:44:51.994323 13551 net.cpp:382] data_data_0_split -> data_data_0_split_2 -I0819 13:44:51.994341 13551 net.cpp:124] Setting up data_data_0_split -I0819 13:44:51.994350 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:51.994361 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:51.994372 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:51.994382 13551 net.cpp:139] Memory required for data: 1600 -I0819 13:44:51.994390 13551 layer_factory.hpp:77] Creating layer NoLevel -I0819 13:44:51.994405 13551 net.cpp:86] Creating Layer NoLevel -I0819 13:44:51.994412 13551 net.cpp:408] NoLevel <- data_data_0_split_0 -I0819 13:44:51.994426 13551 net.cpp:382] NoLevel -> NoLevel -I0819 13:44:51.994459 13551 net.cpp:124] Setting up NoLevel -I0819 13:44:51.994472 13551 net.cpp:131] Top shape: 1 1 (1) -I0819 13:44:51.994482 13551 net.cpp:139] Memory required for data: 1604 -I0819 13:44:51.994505 13551 layer_factory.hpp:77] Creating layer Level0Only -I0819 13:44:51.994525 13551 net.cpp:86] Creating Layer Level0Only -I0819 13:44:51.994537 13551 net.cpp:408] Level0Only <- data_data_0_split_1 -I0819 13:44:51.994555 13551 net.cpp:382] Level0Only -> Level0Only -I0819 13:44:51.994591 13551 net.cpp:124] Setting up Level0Only -I0819 13:44:51.994604 13551 net.cpp:131] Top shape: 1 1 (1) -I0819 13:44:51.994616 13551 net.cpp:139] Memory required for data: 1608 -I0819 13:44:51.994633 13551 layer_factory.hpp:77] Creating layer Level>=0 -I0819 13:44:51.994652 13551 net.cpp:86] Creating Layer Level>=0 -I0819 13:44:51.994664 13551 net.cpp:408] Level>=0 <- data_data_0_split_2 -I0819 13:44:51.994680 13551 net.cpp:382] Level>=0 -> Level>=0 -I0819 13:44:51.994714 13551 net.cpp:124] Setting up Level>=0 -I0819 13:44:51.994724 13551 net.cpp:131] Top shape: 1 1 (1) -I0819 13:44:51.994736 13551 net.cpp:139] Memory required for data: 1612 -I0819 13:44:51.994758 13551 net.cpp:202] Level>=0 does not need backward computation. -I0819 13:44:51.994771 13551 net.cpp:202] Level0Only does not need backward computation. -I0819 13:44:51.994779 13551 net.cpp:202] NoLevel does not need backward computation. -I0819 13:44:51.994788 13551 net.cpp:202] data_data_0_split does not need backward computation. -I0819 13:44:51.994796 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:51.994803 13551 net.cpp:244] This network produces output Level0Only -I0819 13:44:51.994813 13551 net.cpp:244] This network produces output Level>=0 -I0819 13:44:51.994822 13551 net.cpp:244] This network produces output NoLevel -I0819 13:44:51.994840 13551 net.cpp:257] Network initialization done. -.I0819 13:44:51.996243 13551 net.cpp:316] The NetState level (1) is above the max_level (0) specified by a rule in layer Level0Only -I0819 13:44:51.996271 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.383698 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.383726 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.383741 31304 net.cpp:382] data -> data +I0718 09:46:46.383777 31304 net.cpp:124] Setting up data +I0718 09:46:46.383790 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.383803 31304 net.cpp:139] Memory required for data: 400 +I0718 09:46:46.383814 31304 layer_factory.hpp:77] Creating layer data_data_0_split +I0718 09:46:46.383836 31304 net.cpp:86] Creating Layer data_data_0_split +I0718 09:46:46.383846 31304 net.cpp:408] data_data_0_split <- data +I0718 09:46:46.383862 31304 net.cpp:382] data_data_0_split -> data_data_0_split_0 +I0718 09:46:46.383879 31304 net.cpp:382] data_data_0_split -> data_data_0_split_1 +I0718 09:46:46.383901 31304 net.cpp:382] data_data_0_split -> data_data_0_split_2 +I0718 09:46:46.383927 31304 net.cpp:124] Setting up data_data_0_split +I0718 09:46:46.383939 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.383950 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.383960 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.383970 31304 net.cpp:139] Memory required for data: 1600 +I0718 09:46:46.383977 31304 layer_factory.hpp:77] Creating layer NoLevel +I0718 09:46:46.383991 31304 net.cpp:86] Creating Layer NoLevel +I0718 09:46:46.384001 31304 net.cpp:408] NoLevel <- data_data_0_split_0 +I0718 09:46:46.384016 31304 net.cpp:382] NoLevel -> NoLevel +I0718 09:46:46.384052 31304 net.cpp:124] Setting up NoLevel +I0718 09:46:46.384064 31304 net.cpp:131] Top shape: 1 1 (1) +I0718 09:46:46.384078 31304 net.cpp:139] Memory required for data: 1604 +I0718 09:46:46.384100 31304 layer_factory.hpp:77] Creating layer Level0Only +I0718 09:46:46.384125 31304 net.cpp:86] Creating Layer Level0Only +I0718 09:46:46.384136 31304 net.cpp:408] Level0Only <- data_data_0_split_1 +I0718 09:46:46.384153 31304 net.cpp:382] Level0Only -> Level0Only +I0718 09:46:46.384187 31304 net.cpp:124] Setting up Level0Only +I0718 09:46:46.384200 31304 net.cpp:131] Top shape: 1 1 (1) +I0718 09:46:46.384213 31304 net.cpp:139] Memory required for data: 1608 +I0718 09:46:46.384233 31304 layer_factory.hpp:77] Creating layer Level>=0 +I0718 09:46:46.384254 31304 net.cpp:86] Creating Layer Level>=0 +I0718 09:46:46.384264 31304 net.cpp:408] Level>=0 <- data_data_0_split_2 +I0718 09:46:46.384280 31304 net.cpp:382] Level>=0 -> Level>=0 +I0718 09:46:46.384311 31304 net.cpp:124] Setting up Level>=0 +I0718 09:46:46.384321 31304 net.cpp:131] Top shape: 1 1 (1) +I0718 09:46:46.384332 31304 net.cpp:139] Memory required for data: 1612 +I0718 09:46:46.384351 31304 net.cpp:202] Level>=0 does not need backward computation. +I0718 09:46:46.384362 31304 net.cpp:202] Level0Only does not need backward computation. +I0718 09:46:46.384371 31304 net.cpp:202] NoLevel does not need backward computation. +I0718 09:46:46.384382 31304 net.cpp:202] data_data_0_split does not need backward computation. +I0718 09:46:46.384392 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.384402 31304 net.cpp:244] This network produces output Level0Only +I0718 09:46:46.384413 31304 net.cpp:244] This network produces output Level>=0 +I0718 09:46:46.384423 31304 net.cpp:244] This network produces output NoLevel +I0718 09:46:46.384441 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.385887 31304 net.cpp:316] The NetState level (1) is above the max_level (0) specified by a rule in layer Level0Only +I0718 09:46:46.385921 31304 net.cpp:53] Initializing net from parameters: state { phase: TEST level: 1 @@ -15228,65 +15295,65 @@ num_output: 1 } } -I0819 13:44:51.996486 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:51.996510 13551 net.cpp:86] Creating Layer data -I0819 13:44:51.996526 13551 net.cpp:382] data -> data -I0819 13:44:51.996558 13551 net.cpp:124] Setting up data -I0819 13:44:51.996569 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:51.996580 13551 net.cpp:139] Memory required for data: 400 -I0819 13:44:51.996589 13551 layer_factory.hpp:77] Creating layer data_data_0_split -I0819 13:44:51.996604 13551 net.cpp:86] Creating Layer data_data_0_split -I0819 13:44:51.996613 13551 net.cpp:408] data_data_0_split <- data -I0819 13:44:51.996626 13551 net.cpp:382] data_data_0_split -> data_data_0_split_0 -I0819 13:44:51.996642 13551 net.cpp:382] data_data_0_split -> data_data_0_split_1 -I0819 13:44:51.996657 13551 net.cpp:382] data_data_0_split -> data_data_0_split_2 -I0819 13:44:51.996672 13551 net.cpp:382] data_data_0_split -> data_data_0_split_3 -I0819 13:44:51.996691 13551 net.cpp:124] Setting up data_data_0_split -I0819 13:44:51.996701 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:51.996712 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:51.996722 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:51.996732 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:51.996742 13551 net.cpp:139] Memory required for data: 2000 -I0819 13:44:51.996749 13551 layer_factory.hpp:77] Creating layer NoLevel -I0819 13:44:51.996763 13551 net.cpp:86] Creating Layer NoLevel -I0819 13:44:51.996771 13551 net.cpp:408] NoLevel <- data_data_0_split_0 -I0819 13:44:51.996784 13551 net.cpp:382] NoLevel -> NoLevel -I0819 13:44:51.996814 13551 net.cpp:124] Setting up NoLevel -I0819 13:44:51.996824 13551 net.cpp:131] Top shape: 1 1 (1) -I0819 13:44:51.996834 13551 net.cpp:139] Memory required for data: 2004 -I0819 13:44:51.996855 13551 layer_factory.hpp:77] Creating layer Level1Only -I0819 13:44:51.996870 13551 net.cpp:86] Creating Layer Level1Only -I0819 13:44:51.996878 13551 net.cpp:408] Level1Only <- data_data_0_split_1 -I0819 13:44:51.996892 13551 net.cpp:382] Level1Only -> Level1Only -I0819 13:44:51.996918 13551 net.cpp:124] Setting up Level1Only -I0819 13:44:51.996928 13551 net.cpp:131] Top shape: 1 1 (1) -I0819 13:44:51.996937 13551 net.cpp:139] Memory required for data: 2008 -I0819 13:44:51.996953 13551 layer_factory.hpp:77] Creating layer Level>=0 -I0819 13:44:51.996968 13551 net.cpp:86] Creating Layer Level>=0 -I0819 13:44:51.996976 13551 net.cpp:408] Level>=0 <- data_data_0_split_2 -I0819 13:44:51.996989 13551 net.cpp:382] Level>=0 -> Level>=0 -I0819 13:44:51.997014 13551 net.cpp:124] Setting up Level>=0 -I0819 13:44:51.997025 13551 net.cpp:131] Top shape: 1 1 (1) -I0819 13:44:51.997033 13551 net.cpp:139] Memory required for data: 2012 -I0819 13:44:51.997051 13551 layer_factory.hpp:77] Creating layer Level>=1 -I0819 13:44:51.997071 13551 net.cpp:86] Creating Layer Level>=1 -I0819 13:44:51.997081 13551 net.cpp:408] Level>=1 <- data_data_0_split_3 -I0819 13:44:51.997095 13551 net.cpp:382] Level>=1 -> Level>=1 -I0819 13:44:51.997124 13551 net.cpp:124] Setting up Level>=1 -I0819 13:44:51.997135 13551 net.cpp:131] Top shape: 1 1 (1) -I0819 13:44:51.997145 13551 net.cpp:139] Memory required for data: 2016 -I0819 13:44:51.997159 13551 net.cpp:202] Level>=1 does not need backward computation. -I0819 13:44:51.997169 13551 net.cpp:202] Level>=0 does not need backward computation. -I0819 13:44:51.997176 13551 net.cpp:202] Level1Only does not need backward computation. -I0819 13:44:51.997185 13551 net.cpp:202] NoLevel does not need backward computation. -I0819 13:44:51.997193 13551 net.cpp:202] data_data_0_split does not need backward computation. -I0819 13:44:51.997202 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:51.997210 13551 net.cpp:244] This network produces output Level1Only -I0819 13:44:51.997218 13551 net.cpp:244] This network produces output Level>=0 -I0819 13:44:51.997226 13551 net.cpp:244] This network produces output Level>=1 -I0819 13:44:51.997234 13551 net.cpp:244] This network produces output NoLevel -I0819 13:44:51.997251 13551 net.cpp:257] Network initialization done. -.I0819 13:44:51.999814 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.386157 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.386188 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.386204 31304 net.cpp:382] data -> data +I0718 09:46:46.386246 31304 net.cpp:124] Setting up data +I0718 09:46:46.386260 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.386274 31304 net.cpp:139] Memory required for data: 400 +I0718 09:46:46.386286 31304 layer_factory.hpp:77] Creating layer data_data_0_split +I0718 09:46:46.386304 31304 net.cpp:86] Creating Layer data_data_0_split +I0718 09:46:46.386318 31304 net.cpp:408] data_data_0_split <- data +I0718 09:46:46.386338 31304 net.cpp:382] data_data_0_split -> data_data_0_split_0 +I0718 09:46:46.386363 31304 net.cpp:382] data_data_0_split -> data_data_0_split_1 +I0718 09:46:46.386384 31304 net.cpp:382] data_data_0_split -> data_data_0_split_2 +I0718 09:46:46.386404 31304 net.cpp:382] data_data_0_split -> data_data_0_split_3 +I0718 09:46:46.386431 31304 net.cpp:124] Setting up data_data_0_split +I0718 09:46:46.386445 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.386459 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.386471 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.386482 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.386492 31304 net.cpp:139] Memory required for data: 2000 +I0718 09:46:46.386502 31304 layer_factory.hpp:77] Creating layer NoLevel +I0718 09:46:46.386523 31304 net.cpp:86] Creating Layer NoLevel +I0718 09:46:46.386538 31304 net.cpp:408] NoLevel <- data_data_0_split_0 +I0718 09:46:46.386559 31304 net.cpp:382] NoLevel -> NoLevel +I0718 09:46:46.386601 31304 net.cpp:124] Setting up NoLevel +I0718 09:46:46.386613 31304 net.cpp:131] Top shape: 1 1 (1) +I0718 09:46:46.386626 31304 net.cpp:139] Memory required for data: 2004 +I0718 09:46:46.386651 31304 layer_factory.hpp:77] Creating layer Level1Only +I0718 09:46:46.386670 31304 net.cpp:86] Creating Layer Level1Only +I0718 09:46:46.386684 31304 net.cpp:408] Level1Only <- data_data_0_split_1 +I0718 09:46:46.386703 31304 net.cpp:382] Level1Only -> Level1Only +I0718 09:46:46.386736 31304 net.cpp:124] Setting up Level1Only +I0718 09:46:46.386749 31304 net.cpp:131] Top shape: 1 1 (1) +I0718 09:46:46.386761 31304 net.cpp:139] Memory required for data: 2008 +I0718 09:46:46.386780 31304 layer_factory.hpp:77] Creating layer Level>=0 +I0718 09:46:46.386801 31304 net.cpp:86] Creating Layer Level>=0 +I0718 09:46:46.386813 31304 net.cpp:408] Level>=0 <- data_data_0_split_2 +I0718 09:46:46.386831 31304 net.cpp:382] Level>=0 -> Level>=0 +I0718 09:46:46.386864 31304 net.cpp:124] Setting up Level>=0 +I0718 09:46:46.386876 31304 net.cpp:131] Top shape: 1 1 (1) +I0718 09:46:46.386888 31304 net.cpp:139] Memory required for data: 2012 +I0718 09:46:46.386910 31304 layer_factory.hpp:77] Creating layer Level>=1 +I0718 09:46:46.386929 31304 net.cpp:86] Creating Layer Level>=1 +I0718 09:46:46.386940 31304 net.cpp:408] Level>=1 <- data_data_0_split_3 +I0718 09:46:46.386955 31304 net.cpp:382] Level>=1 -> Level>=1 +I0718 09:46:46.386996 31304 net.cpp:124] Setting up Level>=1 +I0718 09:46:46.387009 31304 net.cpp:131] Top shape: 1 1 (1) +I0718 09:46:46.387022 31304 net.cpp:139] Memory required for data: 2016 +I0718 09:46:46.387037 31304 net.cpp:202] Level>=1 does not need backward computation. +I0718 09:46:46.387050 31304 net.cpp:202] Level>=0 does not need backward computation. +I0718 09:46:46.387061 31304 net.cpp:202] Level1Only does not need backward computation. +I0718 09:46:46.387073 31304 net.cpp:202] NoLevel does not need backward computation. +I0718 09:46:46.387084 31304 net.cpp:202] data_data_0_split does not need backward computation. +I0718 09:46:46.387094 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.387100 31304 net.cpp:244] This network produces output Level1Only +I0718 09:46:46.387110 31304 net.cpp:244] This network produces output Level>=0 +I0718 09:46:46.387120 31304 net.cpp:244] This network produces output Level>=1 +I0718 09:46:46.387130 31304 net.cpp:244] This network produces output NoLevel +I0718 09:46:46.387147 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.389035 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -15365,45 +15432,45 @@ bottom: "label" top: "loss" } -I0819 13:44:52.000154 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.000185 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.000202 13551 net.cpp:382] data -> data -I0819 13:44:52.000231 13551 net.cpp:382] data -> label -I0819 13:44:52.000268 13551 net.cpp:124] Setting up data -I0819 13:44:52.000280 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.000295 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.000306 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.000317 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.000344 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.000355 13551 net.cpp:408] conv <- data -I0819 13:44:52.000371 13551 net.cpp:382] conv -> conv -I0819 13:44:52.000547 13551 net.cpp:124] Setting up conv -I0819 13:44:52.000564 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.000577 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.000602 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.000620 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.000630 13551 net.cpp:408] ip <- conv -I0819 13:44:52.000645 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.000933 13551 net.cpp:124] Setting up ip -I0819 13:44:52.000952 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.000963 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.000982 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.000998 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.001008 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.001019 13551 net.cpp:408] loss <- label -I0819 13:44:52.001034 13551 net.cpp:382] loss -> loss -I0819 13:44:52.001058 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.001097 13551 net.cpp:124] Setting up loss -I0819 13:44:52.001107 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.001118 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.001138 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.001147 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.001158 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.001168 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.001178 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.001188 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.001202 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.007006 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.389344 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.389377 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.389396 31304 net.cpp:382] data -> data +I0718 09:46:46.389425 31304 net.cpp:382] data -> label +I0718 09:46:46.389461 31304 net.cpp:124] Setting up data +I0718 09:46:46.389478 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.389494 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.389506 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.389518 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.389552 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.389565 31304 net.cpp:408] conv <- data +I0718 09:46:46.389587 31304 net.cpp:382] conv -> conv +I0718 09:46:46.389786 31304 net.cpp:124] Setting up conv +I0718 09:46:46.389806 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.389817 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.389840 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.389861 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.389873 31304 net.cpp:408] ip <- conv +I0718 09:46:46.389889 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.390159 31304 net.cpp:124] Setting up ip +I0718 09:46:46.390179 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.390192 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.390211 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.390228 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.390239 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.390249 31304 net.cpp:408] loss <- label +I0718 09:46:46.390262 31304 net.cpp:382] loss -> loss +I0718 09:46:46.390286 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.390324 31304 net.cpp:124] Setting up loss +I0718 09:46:46.390334 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.390345 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.390367 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.390377 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.390388 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.390398 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.390408 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.390416 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.390432 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.395790 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -15482,45 +15549,45 @@ bottom: "label" top: "loss" } -I0819 13:44:52.007331 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.007366 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.007385 13551 net.cpp:382] data -> data -I0819 13:44:52.007419 13551 net.cpp:382] data -> label -I0819 13:44:52.007462 13551 net.cpp:124] Setting up data -I0819 13:44:52.009997 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.010062 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.010076 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.010089 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.010133 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.010149 13551 net.cpp:408] conv <- data -I0819 13:44:52.010174 13551 net.cpp:382] conv -> conv -I0819 13:44:52.010267 13551 net.cpp:124] Setting up conv -I0819 13:44:52.010282 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.010295 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.010322 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.010344 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.010355 13551 net.cpp:408] ip <- conv -I0819 13:44:52.010373 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.010648 13551 net.cpp:124] Setting up ip -I0819 13:44:52.010668 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.010679 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.010697 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.010715 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.010723 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.010733 13551 net.cpp:408] loss <- label -I0819 13:44:52.010747 13551 net.cpp:382] loss -> loss -I0819 13:44:52.010772 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.010810 13551 net.cpp:124] Setting up loss -I0819 13:44:52.010821 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.010831 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.010856 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.010865 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.010876 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.010885 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.010893 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.010903 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.010918 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.014636 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.396108 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.396140 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.396159 31304 net.cpp:382] data -> data +I0718 09:46:46.396193 31304 net.cpp:382] data -> label +I0718 09:46:46.396231 31304 net.cpp:124] Setting up data +I0718 09:46:46.396245 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.396261 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.396272 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.396282 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.396308 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.396319 31304 net.cpp:408] conv <- data +I0718 09:46:46.396338 31304 net.cpp:382] conv -> conv +I0718 09:46:46.396407 31304 net.cpp:124] Setting up conv +I0718 09:46:46.396420 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.396430 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.396452 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.396471 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.396483 31304 net.cpp:408] ip <- conv +I0718 09:46:46.396500 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.396775 31304 net.cpp:124] Setting up ip +I0718 09:46:46.396795 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.396806 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.396824 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.396842 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.396850 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.396862 31304 net.cpp:408] loss <- label +I0718 09:46:46.396873 31304 net.cpp:382] loss -> loss +I0718 09:46:46.396899 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.396937 31304 net.cpp:124] Setting up loss +I0718 09:46:46.396947 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.396957 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.396977 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.396986 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.396996 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.397004 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.397014 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.397023 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.397037 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.406790 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -15599,45 +15666,45 @@ bottom: "label" top: "loss" } -I0819 13:44:52.014941 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.014973 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.014992 13551 net.cpp:382] data -> data -I0819 13:44:52.015022 13551 net.cpp:382] data -> label -I0819 13:44:52.015058 13551 net.cpp:124] Setting up data -I0819 13:44:52.015069 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.015084 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.015094 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.015102 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.015126 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.015137 13551 net.cpp:408] conv <- data -I0819 13:44:52.015153 13551 net.cpp:382] conv -> conv -I0819 13:44:52.015213 13551 net.cpp:124] Setting up conv -I0819 13:44:52.015225 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.015237 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.015257 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.015276 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.015286 13551 net.cpp:408] ip <- conv -I0819 13:44:52.015301 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.015563 13551 net.cpp:124] Setting up ip -I0819 13:44:52.015578 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.015588 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.015604 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.015620 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.015630 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.015640 13551 net.cpp:408] loss <- label -I0819 13:44:52.015653 13551 net.cpp:382] loss -> loss -I0819 13:44:52.015676 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.015709 13551 net.cpp:124] Setting up loss -I0819 13:44:52.015720 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.015730 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.015748 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.015756 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.015767 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.015776 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.015785 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.015794 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.015808 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.018683 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.407120 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.407155 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.407173 31304 net.cpp:382] data -> data +I0718 09:46:46.407231 31304 net.cpp:382] data -> label +I0718 09:46:46.407276 31304 net.cpp:124] Setting up data +I0718 09:46:46.407291 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.407310 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.407322 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.407332 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.407359 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.407371 31304 net.cpp:408] conv <- data +I0718 09:46:46.407392 31304 net.cpp:382] conv -> conv +I0718 09:46:46.407476 31304 net.cpp:124] Setting up conv +I0718 09:46:46.407492 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.407506 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.407531 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.407552 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.407564 31304 net.cpp:408] ip <- conv +I0718 09:46:46.407582 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.407863 31304 net.cpp:124] Setting up ip +I0718 09:46:46.407883 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.407896 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.407915 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.407932 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.407941 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.407953 31304 net.cpp:408] loss <- label +I0718 09:46:46.407966 31304 net.cpp:382] loss -> loss +I0718 09:46:46.407994 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.408032 31304 net.cpp:124] Setting up loss +I0718 09:46:46.408046 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.408057 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.408082 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.408092 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.408103 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.408113 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.408123 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.408134 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.408151 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.410956 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -15716,45 +15783,45 @@ bottom: "label" top: "loss" } -I0819 13:44:52.018976 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.019008 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.019026 13551 net.cpp:382] data -> data -I0819 13:44:52.019054 13551 net.cpp:382] data -> label -I0819 13:44:52.019093 13551 net.cpp:124] Setting up data -I0819 13:44:52.019109 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.019124 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.019135 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.019145 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.019174 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.019187 13551 net.cpp:408] conv <- data -I0819 13:44:52.019207 13551 net.cpp:382] conv -> conv -I0819 13:44:52.019276 13551 net.cpp:124] Setting up conv -I0819 13:44:52.019294 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.019306 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.019330 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.019349 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.019359 13551 net.cpp:408] ip <- conv -I0819 13:44:52.019376 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.019644 13551 net.cpp:124] Setting up ip -I0819 13:44:52.019662 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.019675 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.019693 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.019709 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.019718 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.019728 13551 net.cpp:408] loss <- label -I0819 13:44:52.019742 13551 net.cpp:382] loss -> loss -I0819 13:44:52.019764 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.019796 13551 net.cpp:124] Setting up loss -I0819 13:44:52.019805 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.019815 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.019834 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.019842 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.019852 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.019860 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.019870 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.019878 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.019891 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.024085 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.411389 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.411432 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.411449 31304 net.cpp:382] data -> data +I0718 09:46:46.411476 31304 net.cpp:382] data -> label +I0718 09:46:46.411511 31304 net.cpp:124] Setting up data +I0718 09:46:46.411522 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.411537 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.411548 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.411558 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.411585 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.411597 31304 net.cpp:408] conv <- data +I0718 09:46:46.411618 31304 net.cpp:382] conv -> conv +I0718 09:46:46.411687 31304 net.cpp:124] Setting up conv +I0718 09:46:46.411703 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.411717 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.411741 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.411763 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.411774 31304 net.cpp:408] ip <- conv +I0718 09:46:46.411792 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.412072 31304 net.cpp:124] Setting up ip +I0718 09:46:46.412094 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.412107 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.412128 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.412145 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.412155 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.412166 31304 net.cpp:408] loss <- label +I0718 09:46:46.412180 31304 net.cpp:382] loss -> loss +I0718 09:46:46.412210 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.412254 31304 net.cpp:124] Setting up loss +I0718 09:46:46.412271 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.412286 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.412310 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.412322 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.412335 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.412348 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.412360 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.412371 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.412391 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.416774 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -15833,45 +15900,45 @@ bottom: "label" top: "loss" } -I0819 13:44:52.024359 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.024391 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.024410 13551 net.cpp:382] data -> data -I0819 13:44:52.024439 13551 net.cpp:382] data -> label -I0819 13:44:52.024473 13551 net.cpp:124] Setting up data -I0819 13:44:52.024485 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.024502 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.024511 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.024519 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.024546 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.024557 13551 net.cpp:408] conv <- data -I0819 13:44:52.024574 13551 net.cpp:382] conv -> conv -I0819 13:44:52.024638 13551 net.cpp:124] Setting up conv -I0819 13:44:52.024653 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.024665 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.024689 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.024708 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.024720 13551 net.cpp:408] ip <- conv -I0819 13:44:52.024736 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.025004 13551 net.cpp:124] Setting up ip -I0819 13:44:52.025023 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.025035 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.025054 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.025071 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.025080 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.025091 13551 net.cpp:408] loss <- label -I0819 13:44:52.025104 13551 net.cpp:382] loss -> loss -I0819 13:44:52.025128 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.025164 13551 net.cpp:124] Setting up loss -I0819 13:44:52.025177 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.025189 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.025210 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.025220 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.025233 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.025243 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.025252 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.025262 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.025278 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.027376 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.417094 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.417129 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.417150 31304 net.cpp:382] data -> data +I0718 09:46:46.417181 31304 net.cpp:382] data -> label +I0718 09:46:46.417222 31304 net.cpp:124] Setting up data +I0718 09:46:46.417237 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.417254 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.417265 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.417274 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.417304 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.417317 31304 net.cpp:408] conv <- data +I0718 09:46:46.417338 31304 net.cpp:382] conv -> conv +I0718 09:46:46.417419 31304 net.cpp:124] Setting up conv +I0718 09:46:46.417435 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.417449 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.417474 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.417497 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.417507 31304 net.cpp:408] ip <- conv +I0718 09:46:46.417524 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.417807 31304 net.cpp:124] Setting up ip +I0718 09:46:46.417830 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.417842 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.417863 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.417883 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.417893 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.417904 31304 net.cpp:408] loss <- label +I0718 09:46:46.417918 31304 net.cpp:382] loss -> loss +I0718 09:46:46.417948 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.417989 31304 net.cpp:124] Setting up loss +I0718 09:46:46.418001 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.418013 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.418037 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.418047 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.418061 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.418069 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.418079 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.418090 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.418107 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.420425 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -15950,45 +16017,45 @@ bottom: "label" top: "loss" } -I0819 13:44:52.027626 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.027652 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.027669 13551 net.cpp:382] data -> data -I0819 13:44:52.027694 13551 net.cpp:382] data -> label -I0819 13:44:52.027725 13551 net.cpp:124] Setting up data -I0819 13:44:52.027736 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.027752 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.027762 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.027771 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.027796 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.027806 13551 net.cpp:408] conv <- data -I0819 13:44:52.027824 13551 net.cpp:382] conv -> conv -I0819 13:44:52.027889 13551 net.cpp:124] Setting up conv -I0819 13:44:52.027902 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.027913 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.027935 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.027952 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.027962 13551 net.cpp:408] ip <- conv -I0819 13:44:52.027977 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.028244 13551 net.cpp:124] Setting up ip -I0819 13:44:52.028264 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.028276 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.028295 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.028311 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.028321 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.028332 13551 net.cpp:408] loss <- label -I0819 13:44:52.028345 13551 net.cpp:382] loss -> loss -I0819 13:44:52.028368 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.028403 13551 net.cpp:124] Setting up loss -I0819 13:44:52.028414 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.028424 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.028443 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.028452 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.028462 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.028471 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.028481 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.028489 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.028503 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.030807 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.420718 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.420750 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.420768 31304 net.cpp:382] data -> data +I0718 09:46:46.420795 31304 net.cpp:382] data -> label +I0718 09:46:46.420831 31304 net.cpp:124] Setting up data +I0718 09:46:46.420845 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.420858 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.420869 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.420878 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.420905 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.420917 31304 net.cpp:408] conv <- data +I0718 09:46:46.420938 31304 net.cpp:382] conv -> conv +I0718 09:46:46.421011 31304 net.cpp:124] Setting up conv +I0718 09:46:46.421027 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.421041 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.421064 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.421087 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.421097 31304 net.cpp:408] ip <- conv +I0718 09:46:46.421114 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.421393 31304 net.cpp:124] Setting up ip +I0718 09:46:46.421413 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.421425 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.421444 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.421461 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.421471 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.421483 31304 net.cpp:408] loss <- label +I0718 09:46:46.421496 31304 net.cpp:382] loss -> loss +I0718 09:46:46.421523 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.421561 31304 net.cpp:124] Setting up loss +I0718 09:46:46.421573 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.421586 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.421607 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.421617 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.421627 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.421638 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.421648 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.421656 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.421671 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.424023 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -16067,45 +16134,45 @@ bottom: "label" top: "loss" } -I0819 13:44:52.031072 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.031103 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.031121 13551 net.cpp:382] data -> data -I0819 13:44:52.031148 13551 net.cpp:382] data -> label -I0819 13:44:52.031180 13551 net.cpp:124] Setting up data -I0819 13:44:52.031193 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.031209 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.031220 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.031229 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.031256 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.031270 13551 net.cpp:408] conv <- data -I0819 13:44:52.031291 13551 net.cpp:382] conv -> conv -I0819 13:44:52.031359 13551 net.cpp:124] Setting up conv -I0819 13:44:52.031374 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.031386 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.031409 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.031428 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.031440 13551 net.cpp:408] ip <- conv -I0819 13:44:52.031456 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.031726 13551 net.cpp:124] Setting up ip -I0819 13:44:52.031745 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.031757 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.031775 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.031791 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.031801 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.031811 13551 net.cpp:408] loss <- label -I0819 13:44:52.031826 13551 net.cpp:382] loss -> loss -I0819 13:44:52.031850 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.031884 13551 net.cpp:124] Setting up loss -I0819 13:44:52.031898 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.031909 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.031929 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.031939 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.031950 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.031960 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.031970 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.031980 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.031994 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.035321 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.424343 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.424377 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.424396 31304 net.cpp:382] data -> data +I0718 09:46:46.424427 31304 net.cpp:382] data -> label +I0718 09:46:46.424466 31304 net.cpp:124] Setting up data +I0718 09:46:46.424481 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.424499 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.424510 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.424521 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.424551 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.424564 31304 net.cpp:408] conv <- data +I0718 09:46:46.424587 31304 net.cpp:382] conv -> conv +I0718 09:46:46.424662 31304 net.cpp:124] Setting up conv +I0718 09:46:46.424682 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.424696 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.424723 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.424746 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.424759 31304 net.cpp:408] ip <- conv +I0718 09:46:46.424778 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.425060 31304 net.cpp:124] Setting up ip +I0718 09:46:46.425084 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.425098 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.425119 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.425139 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.425151 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.425165 31304 net.cpp:408] loss <- label +I0718 09:46:46.425181 31304 net.cpp:382] loss -> loss +I0718 09:46:46.425213 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.425259 31304 net.cpp:124] Setting up loss +I0718 09:46:46.425276 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.425287 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.425310 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.425320 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.425331 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.425341 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.425352 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.425361 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.425377 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.428570 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -16184,48 +16251,48 @@ bottom: "label" top: "loss" } -I0819 13:44:52.035569 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.035595 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.035612 13551 net.cpp:382] data -> data -I0819 13:44:52.035640 13551 net.cpp:382] data -> label -I0819 13:44:52.035671 13551 net.cpp:124] Setting up data -I0819 13:44:52.035681 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.035696 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.035706 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.035713 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.035737 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.035746 13551 net.cpp:408] conv <- data -I0819 13:44:52.035763 13551 net.cpp:382] conv -> conv -I0819 13:44:52.035820 13551 net.cpp:124] Setting up conv -I0819 13:44:52.035832 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.035843 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.035863 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.035881 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.035889 13551 net.cpp:408] ip <- conv -I0819 13:44:52.035903 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.036166 13551 net.cpp:124] Setting up ip -I0819 13:44:52.036183 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.036195 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.036213 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.036229 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.036238 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.036248 13551 net.cpp:408] loss <- label -I0819 13:44:52.036262 13551 net.cpp:382] loss -> loss -I0819 13:44:52.036283 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.036316 13551 net.cpp:124] Setting up loss -I0819 13:44:52.036326 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.036336 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.036355 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.036363 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.036373 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.036381 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.036391 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.036399 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.036413 13551 net.cpp:257] Network initialization done. -W0819 13:44:52.038666 13551 _caffe.cpp:139] DEPRECATION WARNING - deprecated use of Python interface -W0819 13:44:52.038709 13551 _caffe.cpp:140] Use this instead (with the named "weights" parameter): -W0819 13:44:52.038718 13551 _caffe.cpp:142] Net('/tmp/tmp8rcv_v60', 0, weights='/tmp/tmp4pyqai4j') -I0819 13:44:52.039065 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.428850 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.428879 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.428896 31304 net.cpp:382] data -> data +I0718 09:46:46.428925 31304 net.cpp:382] data -> label +I0718 09:46:46.428962 31304 net.cpp:124] Setting up data +I0718 09:46:46.428974 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.428990 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.429000 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.429010 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.429034 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.429044 31304 net.cpp:408] conv <- data +I0718 09:46:46.429061 31304 net.cpp:382] conv -> conv +I0718 09:46:46.429127 31304 net.cpp:124] Setting up conv +I0718 09:46:46.429138 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.429149 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.429172 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.429188 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.429198 31304 net.cpp:408] ip <- conv +I0718 09:46:46.429212 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.429482 31304 net.cpp:124] Setting up ip +I0718 09:46:46.429503 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.429515 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.429534 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.429553 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.429561 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.429574 31304 net.cpp:408] loss <- label +I0718 09:46:46.429586 31304 net.cpp:382] loss -> loss +I0718 09:46:46.429612 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.429652 31304 net.cpp:124] Setting up loss +I0718 09:46:46.429664 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.429677 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.429697 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.429708 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.429719 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.429728 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.429738 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.429747 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.429764 31304 net.cpp:257] Network initialization done. +W0718 09:46:46.431661 31304 _caffe.cpp:139] DEPRECATION WARNING - deprecated use of Python interface +W0718 09:46:46.431700 31304 _caffe.cpp:140] Use this instead (with the named "weights" parameter): +W0718 09:46:46.431710 31304 _caffe.cpp:142] Net('/tmp/tmp7ebxn7w9', 0, weights='/tmp/tmpbos85z55') +I0718 09:46:46.432050 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -16304,45 +16371,45 @@ bottom: "label" top: "loss" } -I0819 13:44:52.041473 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.041522 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.041543 13551 net.cpp:382] data -> data -I0819 13:44:52.041579 13551 net.cpp:382] data -> label -I0819 13:44:52.041620 13551 net.cpp:124] Setting up data -I0819 13:44:52.041633 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.041651 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.041661 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.041672 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.043534 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.043550 13551 net.cpp:408] conv <- data -I0819 13:44:52.043571 13551 net.cpp:382] conv -> conv -I0819 13:44:52.043664 13551 net.cpp:124] Setting up conv -I0819 13:44:52.043680 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.043697 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.043726 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.043751 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.043764 13551 net.cpp:408] ip <- conv -I0819 13:44:52.043784 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.046469 13551 net.cpp:124] Setting up ip -I0819 13:44:52.046512 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.046526 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.046553 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.046581 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.046594 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.046609 13551 net.cpp:408] loss <- label -I0819 13:44:52.046626 13551 net.cpp:382] loss -> loss -I0819 13:44:52.046658 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.046700 13551 net.cpp:124] Setting up loss -I0819 13:44:52.046710 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.046721 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.046747 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.046756 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.046767 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.046777 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.046787 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.046795 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.046811 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.050112 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.432313 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.432346 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.432366 31304 net.cpp:382] data -> data +I0718 09:46:46.432396 31304 net.cpp:382] data -> label +I0718 09:46:46.432431 31304 net.cpp:124] Setting up data +I0718 09:46:46.432446 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.432461 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.432472 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.432482 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.432509 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.432523 31304 net.cpp:408] conv <- data +I0718 09:46:46.432540 31304 net.cpp:382] conv -> conv +I0718 09:46:46.432610 31304 net.cpp:124] Setting up conv +I0718 09:46:46.432626 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.432639 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.432663 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.432684 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.432696 31304 net.cpp:408] ip <- conv +I0718 09:46:46.432713 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.432988 31304 net.cpp:124] Setting up ip +I0718 09:46:46.433009 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.433022 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.433041 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.433059 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.433069 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.433081 31304 net.cpp:408] loss <- label +I0718 09:46:46.433095 31304 net.cpp:382] loss -> loss +I0718 09:46:46.433118 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.433162 31304 net.cpp:124] Setting up loss +I0718 09:46:46.433176 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.433187 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.433207 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.433218 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.433228 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.433238 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.433248 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.433256 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.433272 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.436384 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -16421,45 +16488,45 @@ bottom: "label" top: "loss" } -I0819 13:44:52.050415 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.050448 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.050468 13551 net.cpp:382] data -> data -I0819 13:44:52.050503 13551 net.cpp:382] data -> label -I0819 13:44:52.050546 13551 net.cpp:124] Setting up data -I0819 13:44:52.050559 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.050577 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.050587 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.050597 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.050626 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.050639 13551 net.cpp:408] conv <- data -I0819 13:44:52.050660 13551 net.cpp:382] conv -> conv -I0819 13:44:52.050737 13551 net.cpp:124] Setting up conv -I0819 13:44:52.050750 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.050760 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.050781 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.050801 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.050810 13551 net.cpp:408] ip <- conv -I0819 13:44:52.050824 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.051100 13551 net.cpp:124] Setting up ip -I0819 13:44:52.051120 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.051132 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.051152 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.051170 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.051179 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.051190 13551 net.cpp:408] loss <- label -I0819 13:44:52.051203 13551 net.cpp:382] loss -> loss -I0819 13:44:52.051226 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.051263 13551 net.cpp:124] Setting up loss -I0819 13:44:52.051273 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.051285 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.051306 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.051316 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.051326 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.051334 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.051344 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.051353 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.051368 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.055464 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.436671 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.436702 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.436719 31304 net.cpp:382] data -> data +I0718 09:46:46.436745 31304 net.cpp:382] data -> label +I0718 09:46:46.436779 31304 net.cpp:124] Setting up data +I0718 09:46:46.436790 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.436806 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.436817 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.436827 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.436852 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.436864 31304 net.cpp:408] conv <- data +I0718 09:46:46.436884 31304 net.cpp:382] conv -> conv +I0718 09:46:46.436950 31304 net.cpp:124] Setting up conv +I0718 09:46:46.436964 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.436976 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.436997 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.437017 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.437028 31304 net.cpp:408] ip <- conv +I0718 09:46:46.437045 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.437315 31304 net.cpp:124] Setting up ip +I0718 09:46:46.437335 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.437345 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.437363 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.437379 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.437389 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.437399 31304 net.cpp:408] loss <- label +I0718 09:46:46.437412 31304 net.cpp:382] loss -> loss +I0718 09:46:46.437433 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.437465 31304 net.cpp:124] Setting up loss +I0718 09:46:46.437475 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.437485 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.437503 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.437511 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.437521 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.437530 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.437539 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.437548 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.437562 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.440613 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -16538,45 +16605,45 @@ bottom: "label" top: "loss" } -I0819 13:44:52.055763 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.055799 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.055815 13551 net.cpp:382] data -> data -I0819 13:44:52.055845 13551 net.cpp:382] data -> label -I0819 13:44:52.055881 13551 net.cpp:124] Setting up data -I0819 13:44:52.055893 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.055909 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.055920 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.055929 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.055953 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.055965 13551 net.cpp:408] conv <- data -I0819 13:44:52.055984 13551 net.cpp:382] conv -> conv -I0819 13:44:52.056054 13551 net.cpp:124] Setting up conv -I0819 13:44:52.056069 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.056082 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.056106 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.056126 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.056136 13551 net.cpp:408] ip <- conv -I0819 13:44:52.056152 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.056434 13551 net.cpp:124] Setting up ip -I0819 13:44:52.056457 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.056469 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.056488 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.056505 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.056515 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.056525 13551 net.cpp:408] loss <- label -I0819 13:44:52.056540 13551 net.cpp:382] loss -> loss -I0819 13:44:52.056565 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.056603 13551 net.cpp:124] Setting up loss -I0819 13:44:52.056615 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.056627 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.056649 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.056659 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.056670 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.056680 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.056690 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.056699 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.056716 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.061491 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.440881 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.440908 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.440924 31304 net.cpp:382] data -> data +I0718 09:46:46.440950 31304 net.cpp:382] data -> label +I0718 09:46:46.440981 31304 net.cpp:124] Setting up data +I0718 09:46:46.440992 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.441009 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.441018 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.441027 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.441049 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.441061 31304 net.cpp:408] conv <- data +I0718 09:46:46.441079 31304 net.cpp:382] conv -> conv +I0718 09:46:46.441145 31304 net.cpp:124] Setting up conv +I0718 09:46:46.441162 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.441176 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.441201 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.441222 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.441232 31304 net.cpp:408] ip <- conv +I0718 09:46:46.441249 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.441514 31304 net.cpp:124] Setting up ip +I0718 09:46:46.441534 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.441545 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.441563 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.441581 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.441592 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.441603 31304 net.cpp:408] loss <- label +I0718 09:46:46.441617 31304 net.cpp:382] loss -> loss +I0718 09:46:46.441639 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.441677 31304 net.cpp:124] Setting up loss +I0718 09:46:46.441689 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.441700 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.441721 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.441731 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.441742 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.441752 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.441762 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.441771 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.441784 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.446090 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -16655,46 +16722,46 @@ bottom: "label" top: "loss" } -I0819 13:44:52.061803 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.061839 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.061857 13551 net.cpp:382] data -> data -I0819 13:44:52.061890 13551 net.cpp:382] data -> label -I0819 13:44:52.061931 13551 net.cpp:124] Setting up data -I0819 13:44:52.061944 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.061960 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.061970 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.061980 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.062005 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.062017 13551 net.cpp:408] conv <- data -I0819 13:44:52.062036 13551 net.cpp:382] conv -> conv -I0819 13:44:52.062108 13551 net.cpp:124] Setting up conv -I0819 13:44:52.062122 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.062135 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.062156 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.062175 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.062184 13551 net.cpp:408] ip <- conv -I0819 13:44:52.062197 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.062474 13551 net.cpp:124] Setting up ip -I0819 13:44:52.062491 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.062503 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.062520 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.062536 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.062544 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.062554 13551 net.cpp:408] loss <- label -I0819 13:44:52.062566 13551 net.cpp:382] loss -> loss -I0819 13:44:52.062588 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.062629 13551 net.cpp:124] Setting up loss -I0819 13:44:52.062642 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.062654 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.062675 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.062685 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.062695 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.062705 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.062714 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.062726 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.062741 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.063663 13551 hdf5.cpp:33] Datatype class: H5T_FLOAT -.I0819 13:44:52.067984 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.446401 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.446439 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.446460 31304 net.cpp:382] data -> data +I0718 09:46:46.446496 31304 net.cpp:382] data -> label +I0718 09:46:46.446540 31304 net.cpp:124] Setting up data +I0718 09:46:46.446554 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.446571 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.446583 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.446594 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.446622 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.446638 31304 net.cpp:408] conv <- data +I0718 09:46:46.446660 31304 net.cpp:382] conv -> conv +I0718 09:46:46.446748 31304 net.cpp:124] Setting up conv +I0718 09:46:46.446770 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.446784 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.446807 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.446826 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.446837 31304 net.cpp:408] ip <- conv +I0718 09:46:46.446854 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.447134 31304 net.cpp:124] Setting up ip +I0718 09:46:46.447155 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.447167 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.447186 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.447443 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.447453 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.447465 31304 net.cpp:408] loss <- label +I0718 09:46:46.447480 31304 net.cpp:382] loss -> loss +I0718 09:46:46.447504 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.447538 31304 net.cpp:124] Setting up loss +I0718 09:46:46.447549 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.447562 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.447582 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.447592 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.447602 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.447611 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.447621 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.447631 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.447647 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.448624 31304 hdf5.cpp:33] Datatype class: H5T_FLOAT +.I0718 09:46:46.453187 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -16773,47 +16840,47 @@ bottom: "label" top: "loss" } -I0819 13:44:52.068274 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.068303 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.068318 13551 net.cpp:382] data -> data -I0819 13:44:52.068344 13551 net.cpp:382] data -> label -I0819 13:44:52.068377 13551 net.cpp:124] Setting up data -I0819 13:44:52.068387 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.068401 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.068411 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.068420 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.068440 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.068450 13551 net.cpp:408] conv <- data -I0819 13:44:52.068465 13551 net.cpp:382] conv -> conv -I0819 13:44:52.068524 13551 net.cpp:124] Setting up conv -I0819 13:44:52.068536 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.068547 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.068568 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.068585 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.068594 13551 net.cpp:408] ip <- conv -I0819 13:44:52.068606 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.068872 13551 net.cpp:124] Setting up ip -I0819 13:44:52.068892 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.068903 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.068923 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.068939 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.068949 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.068960 13551 net.cpp:408] loss <- label -I0819 13:44:52.068974 13551 net.cpp:382] loss -> loss -I0819 13:44:52.068996 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.069032 13551 net.cpp:124] Setting up loss -I0819 13:44:52.069044 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.069056 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.069078 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.069088 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.069098 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.069108 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.069116 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.069126 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.069141 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.071753 13551 net.cpp:332] The NetState did not contain stage 'B' specified by a rule in layer B -I0819 13:44:52.071784 13551 net.cpp:332] The NetState did not contain stage 'B' specified by a rule in layer AandB -I0819 13:44:52.071795 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.453497 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.453529 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.453548 31304 net.cpp:382] data -> data +I0718 09:46:46.453579 31304 net.cpp:382] data -> label +I0718 09:46:46.453621 31304 net.cpp:124] Setting up data +I0718 09:46:46.453636 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.453653 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.453665 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.453676 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.453704 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.453720 31304 net.cpp:408] conv <- data +I0718 09:46:46.453742 31304 net.cpp:382] conv -> conv +I0718 09:46:46.453820 31304 net.cpp:124] Setting up conv +I0718 09:46:46.453841 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.453855 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.453878 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.453899 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.453910 31304 net.cpp:408] ip <- conv +I0718 09:46:46.453927 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.454208 31304 net.cpp:124] Setting up ip +I0718 09:46:46.454231 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.454244 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.454265 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.454283 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.454294 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.454306 31304 net.cpp:408] loss <- label +I0718 09:46:46.454321 31304 net.cpp:382] loss -> loss +I0718 09:46:46.454345 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.454383 31304 net.cpp:124] Setting up loss +I0718 09:46:46.454394 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.454406 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.454427 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.454435 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.454447 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.454455 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.454465 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.454475 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.454492 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.457093 31304 net.cpp:332] The NetState did not contain stage 'B' specified by a rule in layer B +I0718 09:46:46.457129 31304 net.cpp:332] The NetState did not contain stage 'B' specified by a rule in layer AandB +I0718 09:46:46.457139 31304 net.cpp:53] Initializing net from parameters: state { phase: TEST level: 0 @@ -16859,43 +16926,43 @@ num_output: 1 } } -I0819 13:44:52.071990 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.072016 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.072031 13551 net.cpp:382] data -> data -I0819 13:44:52.072067 13551 net.cpp:124] Setting up data -I0819 13:44:52.072079 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:52.072093 13551 net.cpp:139] Memory required for data: 400 -I0819 13:44:52.072103 13551 layer_factory.hpp:77] Creating layer data_data_0_split -I0819 13:44:52.072120 13551 net.cpp:86] Creating Layer data_data_0_split -I0819 13:44:52.072131 13551 net.cpp:408] data_data_0_split <- data -I0819 13:44:52.072146 13551 net.cpp:382] data_data_0_split -> data_data_0_split_0 -I0819 13:44:52.072165 13551 net.cpp:382] data_data_0_split -> data_data_0_split_1 -I0819 13:44:52.072186 13551 net.cpp:124] Setting up data_data_0_split -I0819 13:44:52.072197 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:52.072211 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:52.072221 13551 net.cpp:139] Memory required for data: 1200 -I0819 13:44:52.072230 13551 layer_factory.hpp:77] Creating layer A -I0819 13:44:52.072248 13551 net.cpp:86] Creating Layer A -I0819 13:44:52.072259 13551 net.cpp:408] A <- data_data_0_split_0 -I0819 13:44:52.072275 13551 net.cpp:382] A -> A -I0819 13:44:52.072309 13551 net.cpp:124] Setting up A -I0819 13:44:52.072322 13551 net.cpp:131] Top shape: 1 1 (1) -I0819 13:44:52.072335 13551 net.cpp:139] Memory required for data: 1204 -I0819 13:44:52.072356 13551 layer_factory.hpp:77] Creating layer AorB -I0819 13:44:52.072376 13551 net.cpp:86] Creating Layer AorB -I0819 13:44:52.072386 13551 net.cpp:408] AorB <- data_data_0_split_1 -I0819 13:44:52.072402 13551 net.cpp:382] AorB -> AorB -I0819 13:44:52.072436 13551 net.cpp:124] Setting up AorB -I0819 13:44:52.072451 13551 net.cpp:131] Top shape: 1 1 (1) -I0819 13:44:52.072463 13551 net.cpp:139] Memory required for data: 1208 -I0819 13:44:52.072479 13551 net.cpp:202] AorB does not need backward computation. -I0819 13:44:52.072489 13551 net.cpp:202] A does not need backward computation. -I0819 13:44:52.072499 13551 net.cpp:202] data_data_0_split does not need backward computation. -I0819 13:44:52.072507 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.072515 13551 net.cpp:244] This network produces output A -I0819 13:44:52.072523 13551 net.cpp:244] This network produces output AorB -I0819 13:44:52.072537 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.074092 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.457340 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.457366 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.457381 31304 net.cpp:382] data -> data +I0718 09:46:46.457425 31304 net.cpp:124] Setting up data +I0718 09:46:46.457438 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.457450 31304 net.cpp:139] Memory required for data: 400 +I0718 09:46:46.457459 31304 layer_factory.hpp:77] Creating layer data_data_0_split +I0718 09:46:46.457474 31304 net.cpp:86] Creating Layer data_data_0_split +I0718 09:46:46.457484 31304 net.cpp:408] data_data_0_split <- data +I0718 09:46:46.457500 31304 net.cpp:382] data_data_0_split -> data_data_0_split_0 +I0718 09:46:46.457520 31304 net.cpp:382] data_data_0_split -> data_data_0_split_1 +I0718 09:46:46.457541 31304 net.cpp:124] Setting up data_data_0_split +I0718 09:46:46.457551 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.457563 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.457573 31304 net.cpp:139] Memory required for data: 1200 +I0718 09:46:46.457581 31304 layer_factory.hpp:77] Creating layer A +I0718 09:46:46.457599 31304 net.cpp:86] Creating Layer A +I0718 09:46:46.457612 31304 net.cpp:408] A <- data_data_0_split_0 +I0718 09:46:46.457628 31304 net.cpp:382] A -> A +I0718 09:46:46.457664 31304 net.cpp:124] Setting up A +I0718 09:46:46.457674 31304 net.cpp:131] Top shape: 1 1 (1) +I0718 09:46:46.457686 31304 net.cpp:139] Memory required for data: 1204 +I0718 09:46:46.457708 31304 layer_factory.hpp:77] Creating layer AorB +I0718 09:46:46.457726 31304 net.cpp:86] Creating Layer AorB +I0718 09:46:46.457734 31304 net.cpp:408] AorB <- data_data_0_split_1 +I0718 09:46:46.457752 31304 net.cpp:382] AorB -> AorB +I0718 09:46:46.457782 31304 net.cpp:124] Setting up AorB +I0718 09:46:46.457794 31304 net.cpp:131] Top shape: 1 1 (1) +I0718 09:46:46.457806 31304 net.cpp:139] Memory required for data: 1208 +I0718 09:46:46.457823 31304 net.cpp:202] AorB does not need backward computation. +I0718 09:46:46.457835 31304 net.cpp:202] A does not need backward computation. +I0718 09:46:46.457845 31304 net.cpp:202] data_data_0_split does not need backward computation. +I0718 09:46:46.457855 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.457863 31304 net.cpp:244] This network produces output A +I0718 09:46:46.457871 31304 net.cpp:244] This network produces output AorB +I0718 09:46:46.457888 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.459417 31304 net.cpp:53] Initializing net from parameters: state { phase: TEST level: 0 @@ -16967,68 +17034,68 @@ num_output: 1 } } -I0819 13:44:52.074319 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.074342 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.074358 13551 net.cpp:382] data -> data -I0819 13:44:52.074394 13551 net.cpp:124] Setting up data -I0819 13:44:52.074407 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:52.074420 13551 net.cpp:139] Memory required for data: 400 -I0819 13:44:52.074429 13551 layer_factory.hpp:77] Creating layer data_data_0_split -I0819 13:44:52.074446 13551 net.cpp:86] Creating Layer data_data_0_split -I0819 13:44:52.074457 13551 net.cpp:408] data_data_0_split <- data -I0819 13:44:52.074472 13551 net.cpp:382] data_data_0_split -> data_data_0_split_0 -I0819 13:44:52.074491 13551 net.cpp:382] data_data_0_split -> data_data_0_split_1 -I0819 13:44:52.074508 13551 net.cpp:382] data_data_0_split -> data_data_0_split_2 -I0819 13:44:52.074525 13551 net.cpp:382] data_data_0_split -> data_data_0_split_3 -I0819 13:44:52.074546 13551 net.cpp:124] Setting up data_data_0_split -I0819 13:44:52.074556 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:52.074569 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:52.074580 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:52.074590 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:52.074600 13551 net.cpp:139] Memory required for data: 2000 -I0819 13:44:52.074609 13551 layer_factory.hpp:77] Creating layer A -I0819 13:44:52.074625 13551 net.cpp:86] Creating Layer A -I0819 13:44:52.074635 13551 net.cpp:408] A <- data_data_0_split_0 -I0819 13:44:52.074651 13551 net.cpp:382] A -> A -I0819 13:44:52.074683 13551 net.cpp:124] Setting up A -I0819 13:44:52.074695 13551 net.cpp:131] Top shape: 1 1 (1) -I0819 13:44:52.074707 13551 net.cpp:139] Memory required for data: 2004 -I0819 13:44:52.074728 13551 layer_factory.hpp:77] Creating layer B -I0819 13:44:52.074744 13551 net.cpp:86] Creating Layer B -I0819 13:44:52.074754 13551 net.cpp:408] B <- data_data_0_split_1 -I0819 13:44:52.074771 13551 net.cpp:382] B -> B -I0819 13:44:52.074801 13551 net.cpp:124] Setting up B -I0819 13:44:52.074813 13551 net.cpp:131] Top shape: 1 1 (1) -I0819 13:44:52.074826 13551 net.cpp:139] Memory required for data: 2008 -I0819 13:44:52.074842 13551 layer_factory.hpp:77] Creating layer AorB -I0819 13:44:52.074862 13551 net.cpp:86] Creating Layer AorB -I0819 13:44:52.074873 13551 net.cpp:408] AorB <- data_data_0_split_2 -I0819 13:44:52.074889 13551 net.cpp:382] AorB -> AorB -I0819 13:44:52.074919 13551 net.cpp:124] Setting up AorB -I0819 13:44:52.074930 13551 net.cpp:131] Top shape: 1 1 (1) -I0819 13:44:52.074940 13551 net.cpp:139] Memory required for data: 2012 -I0819 13:44:52.074956 13551 layer_factory.hpp:77] Creating layer AandB -I0819 13:44:52.074971 13551 net.cpp:86] Creating Layer AandB -I0819 13:44:52.074980 13551 net.cpp:408] AandB <- data_data_0_split_3 -I0819 13:44:52.074993 13551 net.cpp:382] AandB -> AandB -I0819 13:44:52.075017 13551 net.cpp:124] Setting up AandB -I0819 13:44:52.075027 13551 net.cpp:131] Top shape: 1 1 (1) -I0819 13:44:52.075037 13551 net.cpp:139] Memory required for data: 2016 -I0819 13:44:52.075050 13551 net.cpp:202] AandB does not need backward computation. -I0819 13:44:52.075062 13551 net.cpp:202] AorB does not need backward computation. -I0819 13:44:52.075070 13551 net.cpp:202] B does not need backward computation. -I0819 13:44:52.075079 13551 net.cpp:202] A does not need backward computation. -I0819 13:44:52.075088 13551 net.cpp:202] data_data_0_split does not need backward computation. -I0819 13:44:52.075096 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.075103 13551 net.cpp:244] This network produces output A -I0819 13:44:52.075111 13551 net.cpp:244] This network produces output AandB -I0819 13:44:52.075120 13551 net.cpp:244] This network produces output AorB -I0819 13:44:52.075129 13551 net.cpp:244] This network produces output B -I0819 13:44:52.075145 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.076490 13551 net.cpp:332] The NetState did not contain stage 'A' specified by a rule in layer A -I0819 13:44:52.076515 13551 net.cpp:332] The NetState did not contain stage 'A' specified by a rule in layer AorB -I0819 13:44:52.076527 13551 net.cpp:332] The NetState did not contain stage 'A' specified by a rule in layer AandB -I0819 13:44:52.076537 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.459679 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.459709 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.459725 31304 net.cpp:382] data -> data +I0718 09:46:46.459765 31304 net.cpp:124] Setting up data +I0718 09:46:46.459779 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.459791 31304 net.cpp:139] Memory required for data: 400 +I0718 09:46:46.459802 31304 layer_factory.hpp:77] Creating layer data_data_0_split +I0718 09:46:46.459820 31304 net.cpp:86] Creating Layer data_data_0_split +I0718 09:46:46.459832 31304 net.cpp:408] data_data_0_split <- data +I0718 09:46:46.459849 31304 net.cpp:382] data_data_0_split -> data_data_0_split_0 +I0718 09:46:46.459870 31304 net.cpp:382] data_data_0_split -> data_data_0_split_1 +I0718 09:46:46.459890 31304 net.cpp:382] data_data_0_split -> data_data_0_split_2 +I0718 09:46:46.459908 31304 net.cpp:382] data_data_0_split -> data_data_0_split_3 +I0718 09:46:46.459931 31304 net.cpp:124] Setting up data_data_0_split +I0718 09:46:46.459944 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.459959 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.459972 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.459986 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.460000 31304 net.cpp:139] Memory required for data: 2000 +I0718 09:46:46.460011 31304 layer_factory.hpp:77] Creating layer A +I0718 09:46:46.460033 31304 net.cpp:86] Creating Layer A +I0718 09:46:46.460047 31304 net.cpp:408] A <- data_data_0_split_0 +I0718 09:46:46.460067 31304 net.cpp:382] A -> A +I0718 09:46:46.460111 31304 net.cpp:124] Setting up A +I0718 09:46:46.460127 31304 net.cpp:131] Top shape: 1 1 (1) +I0718 09:46:46.460142 31304 net.cpp:139] Memory required for data: 2004 +I0718 09:46:46.460167 31304 layer_factory.hpp:77] Creating layer B +I0718 09:46:46.460187 31304 net.cpp:86] Creating Layer B +I0718 09:46:46.460201 31304 net.cpp:408] B <- data_data_0_split_1 +I0718 09:46:46.460219 31304 net.cpp:382] B -> B +I0718 09:46:46.460255 31304 net.cpp:124] Setting up B +I0718 09:46:46.460268 31304 net.cpp:131] Top shape: 1 1 (1) +I0718 09:46:46.460278 31304 net.cpp:139] Memory required for data: 2008 +I0718 09:46:46.460294 31304 layer_factory.hpp:77] Creating layer AorB +I0718 09:46:46.460314 31304 net.cpp:86] Creating Layer AorB +I0718 09:46:46.460323 31304 net.cpp:408] AorB <- data_data_0_split_2 +I0718 09:46:46.460340 31304 net.cpp:382] AorB -> AorB +I0718 09:46:46.460369 31304 net.cpp:124] Setting up AorB +I0718 09:46:46.460381 31304 net.cpp:131] Top shape: 1 1 (1) +I0718 09:46:46.460391 31304 net.cpp:139] Memory required for data: 2012 +I0718 09:46:46.460408 31304 layer_factory.hpp:77] Creating layer AandB +I0718 09:46:46.460426 31304 net.cpp:86] Creating Layer AandB +I0718 09:46:46.460436 31304 net.cpp:408] AandB <- data_data_0_split_3 +I0718 09:46:46.460453 31304 net.cpp:382] AandB -> AandB +I0718 09:46:46.460480 31304 net.cpp:124] Setting up AandB +I0718 09:46:46.460491 31304 net.cpp:131] Top shape: 1 1 (1) +I0718 09:46:46.460501 31304 net.cpp:139] Memory required for data: 2016 +I0718 09:46:46.460515 31304 net.cpp:202] AandB does not need backward computation. +I0718 09:46:46.460527 31304 net.cpp:202] AorB does not need backward computation. +I0718 09:46:46.460536 31304 net.cpp:202] B does not need backward computation. +I0718 09:46:46.460546 31304 net.cpp:202] A does not need backward computation. +I0718 09:46:46.460554 31304 net.cpp:202] data_data_0_split does not need backward computation. +I0718 09:46:46.460563 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.460572 31304 net.cpp:244] This network produces output A +I0718 09:46:46.460580 31304 net.cpp:244] This network produces output AandB +I0718 09:46:46.460590 31304 net.cpp:244] This network produces output AorB +I0718 09:46:46.460598 31304 net.cpp:244] This network produces output B +I0718 09:46:46.460618 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.462118 31304 net.cpp:332] The NetState did not contain stage 'A' specified by a rule in layer A +I0718 09:46:46.462148 31304 net.cpp:332] The NetState did not contain stage 'A' specified by a rule in layer AorB +I0718 09:46:46.462160 31304 net.cpp:332] The NetState did not contain stage 'A' specified by a rule in layer AandB +I0718 09:46:46.462170 31304 net.cpp:53] Initializing net from parameters: state { phase: TEST level: 0 @@ -17074,43 +17141,43 @@ num_output: 1 } } -I0819 13:44:52.076696 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.076718 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.076733 13551 net.cpp:382] data -> data -I0819 13:44:52.076767 13551 net.cpp:124] Setting up data -I0819 13:44:52.076778 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:52.076789 13551 net.cpp:139] Memory required for data: 400 -I0819 13:44:52.076798 13551 layer_factory.hpp:77] Creating layer data_data_0_split -I0819 13:44:52.076812 13551 net.cpp:86] Creating Layer data_data_0_split -I0819 13:44:52.076822 13551 net.cpp:408] data_data_0_split <- data -I0819 13:44:52.076835 13551 net.cpp:382] data_data_0_split -> data_data_0_split_0 -I0819 13:44:52.076851 13551 net.cpp:382] data_data_0_split -> data_data_0_split_1 -I0819 13:44:52.076870 13551 net.cpp:124] Setting up data_data_0_split -I0819 13:44:52.076880 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:52.076892 13551 net.cpp:131] Top shape: 1 1 10 10 (100) -I0819 13:44:52.076905 13551 net.cpp:139] Memory required for data: 1200 -I0819 13:44:52.076912 13551 layer_factory.hpp:77] Creating layer B -I0819 13:44:52.076929 13551 net.cpp:86] Creating Layer B -I0819 13:44:52.076941 13551 net.cpp:408] B <- data_data_0_split_0 -I0819 13:44:52.076956 13551 net.cpp:382] B -> B -I0819 13:44:52.076987 13551 net.cpp:124] Setting up B -I0819 13:44:52.076997 13551 net.cpp:131] Top shape: 1 1 (1) -I0819 13:44:52.077008 13551 net.cpp:139] Memory required for data: 1204 -I0819 13:44:52.077028 13551 layer_factory.hpp:77] Creating layer AorB -I0819 13:44:52.077046 13551 net.cpp:86] Creating Layer AorB -I0819 13:44:52.077057 13551 net.cpp:408] AorB <- data_data_0_split_1 -I0819 13:44:52.077073 13551 net.cpp:382] AorB -> AorB -I0819 13:44:52.077100 13551 net.cpp:124] Setting up AorB -I0819 13:44:52.077111 13551 net.cpp:131] Top shape: 1 1 (1) -I0819 13:44:52.077122 13551 net.cpp:139] Memory required for data: 1208 -I0819 13:44:52.077138 13551 net.cpp:202] AorB does not need backward computation. -I0819 13:44:52.077147 13551 net.cpp:202] B does not need backward computation. -I0819 13:44:52.077157 13551 net.cpp:202] data_data_0_split does not need backward computation. -I0819 13:44:52.077167 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.077174 13551 net.cpp:244] This network produces output AorB -I0819 13:44:52.077185 13551 net.cpp:244] This network produces output B -I0819 13:44:52.077200 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.081694 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.462349 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.462375 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.462391 31304 net.cpp:382] data -> data +I0718 09:46:46.462431 31304 net.cpp:124] Setting up data +I0718 09:46:46.462445 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.462457 31304 net.cpp:139] Memory required for data: 400 +I0718 09:46:46.462467 31304 layer_factory.hpp:77] Creating layer data_data_0_split +I0718 09:46:46.462483 31304 net.cpp:86] Creating Layer data_data_0_split +I0718 09:46:46.462494 31304 net.cpp:408] data_data_0_split <- data +I0718 09:46:46.462512 31304 net.cpp:382] data_data_0_split -> data_data_0_split_0 +I0718 09:46:46.462529 31304 net.cpp:382] data_data_0_split -> data_data_0_split_1 +I0718 09:46:46.462550 31304 net.cpp:124] Setting up data_data_0_split +I0718 09:46:46.462560 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.462574 31304 net.cpp:131] Top shape: 1 1 10 10 (100) +I0718 09:46:46.462584 31304 net.cpp:139] Memory required for data: 1200 +I0718 09:46:46.462592 31304 layer_factory.hpp:77] Creating layer B +I0718 09:46:46.462610 31304 net.cpp:86] Creating Layer B +I0718 09:46:46.462621 31304 net.cpp:408] B <- data_data_0_split_0 +I0718 09:46:46.462638 31304 net.cpp:382] B -> B +I0718 09:46:46.462672 31304 net.cpp:124] Setting up B +I0718 09:46:46.462685 31304 net.cpp:131] Top shape: 1 1 (1) +I0718 09:46:46.462697 31304 net.cpp:139] Memory required for data: 1204 +I0718 09:46:46.462719 31304 layer_factory.hpp:77] Creating layer AorB +I0718 09:46:46.462738 31304 net.cpp:86] Creating Layer AorB +I0718 09:46:46.462749 31304 net.cpp:408] AorB <- data_data_0_split_1 +I0718 09:46:46.462766 31304 net.cpp:382] AorB -> AorB +I0718 09:46:46.462800 31304 net.cpp:124] Setting up AorB +I0718 09:46:46.462813 31304 net.cpp:131] Top shape: 1 1 (1) +I0718 09:46:46.462826 31304 net.cpp:139] Memory required for data: 1208 +I0718 09:46:46.462842 31304 net.cpp:202] AorB does not need backward computation. +I0718 09:46:46.462853 31304 net.cpp:202] B does not need backward computation. +I0718 09:46:46.462863 31304 net.cpp:202] data_data_0_split does not need backward computation. +I0718 09:46:46.462872 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.462880 31304 net.cpp:244] This network produces output AorB +I0718 09:46:46.462891 31304 net.cpp:244] This network produces output B +I0718 09:46:46.462908 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.467320 31304 net.cpp:53] Initializing net from parameters: state { phase: TEST level: 0 @@ -17223,85 +17290,85 @@ bottom: "label" top: "loss" } -I0819 13:44:52.082006 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.082036 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.082052 13551 net.cpp:382] data -> data -I0819 13:44:52.082083 13551 net.cpp:382] data -> label -I0819 13:44:52.082168 13551 net.cpp:124] Setting up data -I0819 13:44:52.082183 13551 net.cpp:131] Top shape: 50 1 28 28 (39200) -I0819 13:44:52.082203 13551 net.cpp:131] Top shape: 50 1 1 1 (50) -I0819 13:44:52.082216 13551 net.cpp:139] Memory required for data: 157000 -I0819 13:44:52.082227 13551 layer_factory.hpp:77] Creating layer conv1 -I0819 13:44:52.082249 13551 net.cpp:86] Creating Layer conv1 -I0819 13:44:52.082262 13551 net.cpp:408] conv1 <- data -I0819 13:44:52.082278 13551 net.cpp:382] conv1 -> conv1 -I0819 13:44:52.082358 13551 net.cpp:124] Setting up conv1 -I0819 13:44:52.082373 13551 net.cpp:131] Top shape: 50 20 24 24 (576000) -I0819 13:44:52.082388 13551 net.cpp:139] Memory required for data: 2461000 -I0819 13:44:52.082412 13551 layer_factory.hpp:77] Creating layer pool1 -I0819 13:44:52.082432 13551 net.cpp:86] Creating Layer pool1 -I0819 13:44:52.082440 13551 net.cpp:408] pool1 <- conv1 -I0819 13:44:52.082456 13551 net.cpp:382] pool1 -> pool1 -I0819 13:44:52.082504 13551 net.cpp:124] Setting up pool1 -I0819 13:44:52.082516 13551 net.cpp:131] Top shape: 50 20 12 12 (144000) -I0819 13:44:52.082531 13551 net.cpp:139] Memory required for data: 3037000 -I0819 13:44:52.082540 13551 layer_factory.hpp:77] Creating layer conv2 -I0819 13:44:52.082561 13551 net.cpp:86] Creating Layer conv2 -I0819 13:44:52.082572 13551 net.cpp:408] conv2 <- pool1 -I0819 13:44:52.082590 13551 net.cpp:382] conv2 -> conv2 -I0819 13:44:52.083024 13551 net.cpp:124] Setting up conv2 -I0819 13:44:52.083045 13551 net.cpp:131] Top shape: 50 50 8 8 (160000) -I0819 13:44:52.083061 13551 net.cpp:139] Memory required for data: 3677000 -I0819 13:44:52.083081 13551 layer_factory.hpp:77] Creating layer pool2 -I0819 13:44:52.083098 13551 net.cpp:86] Creating Layer pool2 -I0819 13:44:52.083108 13551 net.cpp:408] pool2 <- conv2 -I0819 13:44:52.083124 13551 net.cpp:382] pool2 -> pool2 -I0819 13:44:52.083149 13551 net.cpp:124] Setting up pool2 -I0819 13:44:52.083159 13551 net.cpp:131] Top shape: 50 50 4 4 (40000) -I0819 13:44:52.083173 13551 net.cpp:139] Memory required for data: 3837000 -I0819 13:44:52.083181 13551 layer_factory.hpp:77] Creating layer ip1 -I0819 13:44:52.083196 13551 net.cpp:86] Creating Layer ip1 -I0819 13:44:52.083205 13551 net.cpp:408] ip1 <- pool2 -I0819 13:44:52.083220 13551 net.cpp:382] ip1 -> ip1 -I0819 13:44:52.089418 13551 net.cpp:124] Setting up ip1 -I0819 13:44:52.089486 13551 net.cpp:131] Top shape: 50 500 (25000) -I0819 13:44:52.089504 13551 net.cpp:139] Memory required for data: 3937000 -I0819 13:44:52.089531 13551 layer_factory.hpp:77] Creating layer relu1 -I0819 13:44:52.089553 13551 net.cpp:86] Creating Layer relu1 -I0819 13:44:52.089565 13551 net.cpp:408] relu1 <- ip1 -I0819 13:44:52.089581 13551 net.cpp:369] relu1 -> ip1 (in-place) -I0819 13:44:52.089605 13551 net.cpp:124] Setting up relu1 -I0819 13:44:52.089615 13551 net.cpp:131] Top shape: 50 500 (25000) -I0819 13:44:52.089625 13551 net.cpp:139] Memory required for data: 4037000 -I0819 13:44:52.089633 13551 layer_factory.hpp:77] Creating layer ip2 -I0819 13:44:52.089646 13551 net.cpp:86] Creating Layer ip2 -I0819 13:44:52.089655 13551 net.cpp:408] ip2 <- ip1 -I0819 13:44:52.089669 13551 net.cpp:382] ip2 -> ip2 -I0819 13:44:52.089789 13551 net.cpp:124] Setting up ip2 -I0819 13:44:52.089805 13551 net.cpp:131] Top shape: 50 10 (500) -I0819 13:44:52.089818 13551 net.cpp:139] Memory required for data: 4039000 -I0819 13:44:52.089831 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.089849 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.089860 13551 net.cpp:408] loss <- ip2 -I0819 13:44:52.089871 13551 net.cpp:408] loss <- label -I0819 13:44:52.089885 13551 net.cpp:382] loss -> loss -I0819 13:44:52.089911 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.089951 13551 net.cpp:124] Setting up loss -I0819 13:44:52.089962 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.089974 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.090003 13551 net.cpp:139] Memory required for data: 4039004 -I0819 13:44:52.090013 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.090023 13551 net.cpp:200] ip2 needs backward computation. -I0819 13:44:52.090032 13551 net.cpp:200] relu1 needs backward computation. -I0819 13:44:52.090040 13551 net.cpp:200] ip1 needs backward computation. -I0819 13:44:52.090049 13551 net.cpp:200] pool2 needs backward computation. -I0819 13:44:52.090057 13551 net.cpp:200] conv2 needs backward computation. -I0819 13:44:52.090066 13551 net.cpp:200] pool1 needs backward computation. -I0819 13:44:52.090075 13551 net.cpp:200] conv1 needs backward computation. -I0819 13:44:52.090085 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.090092 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.090112 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.093348 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.467651 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.467681 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.467698 31304 net.cpp:382] data -> data +I0718 09:46:46.467732 31304 net.cpp:382] data -> label +I0718 09:46:46.467819 31304 net.cpp:124] Setting up data +I0718 09:46:46.467834 31304 net.cpp:131] Top shape: 50 1 28 28 (39200) +I0718 09:46:46.467855 31304 net.cpp:131] Top shape: 50 1 1 1 (50) +I0718 09:46:46.467867 31304 net.cpp:139] Memory required for data: 157000 +I0718 09:46:46.467880 31304 layer_factory.hpp:77] Creating layer conv1 +I0718 09:46:46.467905 31304 net.cpp:86] Creating Layer conv1 +I0718 09:46:46.467919 31304 net.cpp:408] conv1 <- data +I0718 09:46:46.467939 31304 net.cpp:382] conv1 -> conv1 +I0718 09:46:46.468021 31304 net.cpp:124] Setting up conv1 +I0718 09:46:46.468036 31304 net.cpp:131] Top shape: 50 20 24 24 (576000) +I0718 09:46:46.468051 31304 net.cpp:139] Memory required for data: 2461000 +I0718 09:46:46.468075 31304 layer_factory.hpp:77] Creating layer pool1 +I0718 09:46:46.468093 31304 net.cpp:86] Creating Layer pool1 +I0718 09:46:46.468104 31304 net.cpp:408] pool1 <- conv1 +I0718 09:46:46.468119 31304 net.cpp:382] pool1 -> pool1 +I0718 09:46:46.468166 31304 net.cpp:124] Setting up pool1 +I0718 09:46:46.468179 31304 net.cpp:131] Top shape: 50 20 12 12 (144000) +I0718 09:46:46.468194 31304 net.cpp:139] Memory required for data: 3037000 +I0718 09:46:46.468204 31304 layer_factory.hpp:77] Creating layer conv2 +I0718 09:46:46.468225 31304 net.cpp:86] Creating Layer conv2 +I0718 09:46:46.468236 31304 net.cpp:408] conv2 <- pool1 +I0718 09:46:46.468253 31304 net.cpp:382] conv2 -> conv2 +I0718 09:46:46.468695 31304 net.cpp:124] Setting up conv2 +I0718 09:46:46.468721 31304 net.cpp:131] Top shape: 50 50 8 8 (160000) +I0718 09:46:46.468736 31304 net.cpp:139] Memory required for data: 3677000 +I0718 09:46:46.468755 31304 layer_factory.hpp:77] Creating layer pool2 +I0718 09:46:46.468771 31304 net.cpp:86] Creating Layer pool2 +I0718 09:46:46.468781 31304 net.cpp:408] pool2 <- conv2 +I0718 09:46:46.468796 31304 net.cpp:382] pool2 -> pool2 +I0718 09:46:46.468823 31304 net.cpp:124] Setting up pool2 +I0718 09:46:46.468834 31304 net.cpp:131] Top shape: 50 50 4 4 (40000) +I0718 09:46:46.468847 31304 net.cpp:139] Memory required for data: 3837000 +I0718 09:46:46.468855 31304 layer_factory.hpp:77] Creating layer ip1 +I0718 09:46:46.468871 31304 net.cpp:86] Creating Layer ip1 +I0718 09:46:46.468880 31304 net.cpp:408] ip1 <- pool2 +I0718 09:46:46.468895 31304 net.cpp:382] ip1 -> ip1 +I0718 09:46:46.474901 31304 net.cpp:124] Setting up ip1 +I0718 09:46:46.474958 31304 net.cpp:131] Top shape: 50 500 (25000) +I0718 09:46:46.474977 31304 net.cpp:139] Memory required for data: 3937000 +I0718 09:46:46.475008 31304 layer_factory.hpp:77] Creating layer relu1 +I0718 09:46:46.475033 31304 net.cpp:86] Creating Layer relu1 +I0718 09:46:46.475047 31304 net.cpp:408] relu1 <- ip1 +I0718 09:46:46.475067 31304 net.cpp:369] relu1 -> ip1 (in-place) +I0718 09:46:46.475093 31304 net.cpp:124] Setting up relu1 +I0718 09:46:46.475103 31304 net.cpp:131] Top shape: 50 500 (25000) +I0718 09:46:46.475114 31304 net.cpp:139] Memory required for data: 4037000 +I0718 09:46:46.475122 31304 layer_factory.hpp:77] Creating layer ip2 +I0718 09:46:46.475138 31304 net.cpp:86] Creating Layer ip2 +I0718 09:46:46.475148 31304 net.cpp:408] ip2 <- ip1 +I0718 09:46:46.475163 31304 net.cpp:382] ip2 -> ip2 +I0718 09:46:46.475384 31304 net.cpp:124] Setting up ip2 +I0718 09:46:46.475410 31304 net.cpp:131] Top shape: 50 10 (500) +I0718 09:46:46.475423 31304 net.cpp:139] Memory required for data: 4039000 +I0718 09:46:46.475438 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.475458 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.475468 31304 net.cpp:408] loss <- ip2 +I0718 09:46:46.475479 31304 net.cpp:408] loss <- label +I0718 09:46:46.475493 31304 net.cpp:382] loss -> loss +I0718 09:46:46.475518 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.475556 31304 net.cpp:124] Setting up loss +I0718 09:46:46.475569 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.475579 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.475611 31304 net.cpp:139] Memory required for data: 4039004 +I0718 09:46:46.475620 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.475631 31304 net.cpp:200] ip2 needs backward computation. +I0718 09:46:46.475641 31304 net.cpp:200] relu1 needs backward computation. +I0718 09:46:46.475651 31304 net.cpp:200] ip1 needs backward computation. +I0718 09:46:46.475659 31304 net.cpp:200] pool2 needs backward computation. +I0718 09:46:46.475668 31304 net.cpp:200] conv2 needs backward computation. +I0718 09:46:46.475677 31304 net.cpp:200] pool1 needs backward computation. +I0718 09:46:46.475687 31304 net.cpp:200] conv1 needs backward computation. +I0718 09:46:46.475697 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.475706 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.475728 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.479425 31304 net.cpp:53] Initializing net from parameters: state { phase: TEST level: 0 @@ -17414,87 +17481,87 @@ bottom: "DummyData2" top: "SoftmaxWithLoss1" } -I0819 13:44:52.093983 13551 layer_factory.hpp:77] Creating layer DummyData1 -I0819 13:44:52.094015 13551 net.cpp:86] Creating Layer DummyData1 -I0819 13:44:52.094033 13551 net.cpp:382] DummyData1 -> DummyData1 -I0819 13:44:52.094061 13551 net.cpp:382] DummyData1 -> DummyData2 -I0819 13:44:52.094167 13551 net.cpp:124] Setting up DummyData1 -I0819 13:44:52.094180 13551 net.cpp:131] Top shape: 50 1 28 28 (39200) -I0819 13:44:52.094198 13551 net.cpp:131] Top shape: 50 1 1 1 (50) -I0819 13:44:52.094208 13551 net.cpp:139] Memory required for data: 157000 -I0819 13:44:52.094218 13551 layer_factory.hpp:77] Creating layer Convolution1 -I0819 13:44:52.094238 13551 net.cpp:86] Creating Layer Convolution1 -I0819 13:44:52.094246 13551 net.cpp:408] Convolution1 <- DummyData1 -I0819 13:44:52.094261 13551 net.cpp:382] Convolution1 -> Convolution1 -I0819 13:44:52.094327 13551 net.cpp:124] Setting up Convolution1 -I0819 13:44:52.094341 13551 net.cpp:131] Top shape: 50 20 24 24 (576000) -I0819 13:44:52.094354 13551 net.cpp:139] Memory required for data: 2461000 -I0819 13:44:52.094375 13551 layer_factory.hpp:77] Creating layer Pooling1 -I0819 13:44:52.094398 13551 net.cpp:86] Creating Layer Pooling1 -I0819 13:44:52.094409 13551 net.cpp:408] Pooling1 <- Convolution1 -I0819 13:44:52.094424 13551 net.cpp:382] Pooling1 -> Pooling1 -I0819 13:44:52.094450 13551 net.cpp:124] Setting up Pooling1 -I0819 13:44:52.094462 13551 net.cpp:131] Top shape: 50 20 12 12 (144000) -I0819 13:44:52.094476 13551 net.cpp:139] Memory required for data: 3037000 -I0819 13:44:52.094485 13551 layer_factory.hpp:77] Creating layer Convolution2 -I0819 13:44:52.094502 13551 net.cpp:86] Creating Layer Convolution2 -I0819 13:44:52.094513 13551 net.cpp:408] Convolution2 <- Pooling1 -I0819 13:44:52.094528 13551 net.cpp:382] Convolution2 -> Convolution2 -I0819 13:44:52.094961 13551 net.cpp:124] Setting up Convolution2 -I0819 13:44:52.094982 13551 net.cpp:131] Top shape: 50 50 8 8 (160000) -I0819 13:44:52.094997 13551 net.cpp:139] Memory required for data: 3677000 -I0819 13:44:52.095014 13551 layer_factory.hpp:77] Creating layer Pooling2 -I0819 13:44:52.095038 13551 net.cpp:86] Creating Layer Pooling2 -I0819 13:44:52.095049 13551 net.cpp:408] Pooling2 <- Convolution2 -I0819 13:44:52.095064 13551 net.cpp:382] Pooling2 -> Pooling2 -I0819 13:44:52.095088 13551 net.cpp:124] Setting up Pooling2 -I0819 13:44:52.095100 13551 net.cpp:131] Top shape: 50 50 4 4 (40000) -I0819 13:44:52.095116 13551 net.cpp:139] Memory required for data: 3837000 -I0819 13:44:52.095125 13551 layer_factory.hpp:77] Creating layer InnerProduct1 -I0819 13:44:52.095139 13551 net.cpp:86] Creating Layer InnerProduct1 -I0819 13:44:52.095149 13551 net.cpp:408] InnerProduct1 <- Pooling2 -I0819 13:44:52.095163 13551 net.cpp:382] InnerProduct1 -> InnerProduct1 -I0819 13:44:52.101415 13551 net.cpp:124] Setting up InnerProduct1 -I0819 13:44:52.101491 13551 net.cpp:131] Top shape: 50 500 (25000) -I0819 13:44:52.101509 13551 net.cpp:139] Memory required for data: 3937000 -I0819 13:44:52.101539 13551 layer_factory.hpp:77] Creating layer ReLU1 -I0819 13:44:52.101563 13551 net.cpp:86] Creating Layer ReLU1 -I0819 13:44:52.101577 13551 net.cpp:408] ReLU1 <- InnerProduct1 -I0819 13:44:52.101594 13551 net.cpp:369] ReLU1 -> InnerProduct1 (in-place) -I0819 13:44:52.101619 13551 net.cpp:124] Setting up ReLU1 -I0819 13:44:52.101627 13551 net.cpp:131] Top shape: 50 500 (25000) -I0819 13:44:52.101639 13551 net.cpp:139] Memory required for data: 4037000 -I0819 13:44:52.101645 13551 layer_factory.hpp:77] Creating layer InnerProduct2 -I0819 13:44:52.101660 13551 net.cpp:86] Creating Layer InnerProduct2 -I0819 13:44:52.101670 13551 net.cpp:408] InnerProduct2 <- InnerProduct1 -I0819 13:44:52.101683 13551 net.cpp:382] InnerProduct2 -> InnerProduct2 -I0819 13:44:52.101807 13551 net.cpp:124] Setting up InnerProduct2 -I0819 13:44:52.101822 13551 net.cpp:131] Top shape: 50 10 (500) -I0819 13:44:52.101835 13551 net.cpp:139] Memory required for data: 4039000 -I0819 13:44:52.101850 13551 layer_factory.hpp:77] Creating layer SoftmaxWithLoss1 -I0819 13:44:52.101871 13551 net.cpp:86] Creating Layer SoftmaxWithLoss1 -I0819 13:44:52.101882 13551 net.cpp:408] SoftmaxWithLoss1 <- InnerProduct2 -I0819 13:44:52.101894 13551 net.cpp:408] SoftmaxWithLoss1 <- DummyData2 -I0819 13:44:52.101912 13551 net.cpp:382] SoftmaxWithLoss1 -> SoftmaxWithLoss1 -I0819 13:44:52.101940 13551 layer_factory.hpp:77] Creating layer SoftmaxWithLoss1 -I0819 13:44:52.101981 13551 net.cpp:124] Setting up SoftmaxWithLoss1 -I0819 13:44:52.101994 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.102005 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.102034 13551 net.cpp:139] Memory required for data: 4039004 -I0819 13:44:52.102044 13551 net.cpp:200] SoftmaxWithLoss1 needs backward computation. -I0819 13:44:52.102056 13551 net.cpp:200] InnerProduct2 needs backward computation. -I0819 13:44:52.102066 13551 net.cpp:200] ReLU1 needs backward computation. -I0819 13:44:52.102074 13551 net.cpp:200] InnerProduct1 needs backward computation. -I0819 13:44:52.102082 13551 net.cpp:200] Pooling2 needs backward computation. -I0819 13:44:52.102092 13551 net.cpp:200] Convolution2 needs backward computation. -I0819 13:44:52.102100 13551 net.cpp:200] Pooling1 needs backward computation. -I0819 13:44:52.102109 13551 net.cpp:200] Convolution1 needs backward computation. -I0819 13:44:52.102119 13551 net.cpp:202] DummyData1 does not need backward computation. -I0819 13:44:52.102128 13551 net.cpp:244] This network produces output SoftmaxWithLoss1 -I0819 13:44:52.102149 13551 net.cpp:257] Network initialization done. +I0718 09:46:46.479764 31304 layer_factory.hpp:77] Creating layer DummyData1 +I0718 09:46:46.479800 31304 net.cpp:86] Creating Layer DummyData1 +I0718 09:46:46.479815 31304 net.cpp:382] DummyData1 -> DummyData1 +I0718 09:46:46.479844 31304 net.cpp:382] DummyData1 -> DummyData2 +I0718 09:46:46.479959 31304 net.cpp:124] Setting up DummyData1 +I0718 09:46:46.479971 31304 net.cpp:131] Top shape: 50 1 28 28 (39200) +I0718 09:46:46.479992 31304 net.cpp:131] Top shape: 50 1 1 1 (50) +I0718 09:46:46.480003 31304 net.cpp:139] Memory required for data: 157000 +I0718 09:46:46.480013 31304 layer_factory.hpp:77] Creating layer Convolution1 +I0718 09:46:46.480034 31304 net.cpp:86] Creating Layer Convolution1 +I0718 09:46:46.480046 31304 net.cpp:408] Convolution1 <- DummyData1 +I0718 09:46:46.480064 31304 net.cpp:382] Convolution1 -> Convolution1 +I0718 09:46:46.480144 31304 net.cpp:124] Setting up Convolution1 +I0718 09:46:46.480157 31304 net.cpp:131] Top shape: 50 20 24 24 (576000) +I0718 09:46:46.480173 31304 net.cpp:139] Memory required for data: 2461000 +I0718 09:46:46.480195 31304 layer_factory.hpp:77] Creating layer Pooling1 +I0718 09:46:46.480214 31304 net.cpp:86] Creating Layer Pooling1 +I0718 09:46:46.480224 31304 net.cpp:408] Pooling1 <- Convolution1 +I0718 09:46:46.480238 31304 net.cpp:382] Pooling1 -> Pooling1 +I0718 09:46:46.480265 31304 net.cpp:124] Setting up Pooling1 +I0718 09:46:46.480276 31304 net.cpp:131] Top shape: 50 20 12 12 (144000) +I0718 09:46:46.480289 31304 net.cpp:139] Memory required for data: 3037000 +I0718 09:46:46.480298 31304 layer_factory.hpp:77] Creating layer Convolution2 +I0718 09:46:46.480316 31304 net.cpp:86] Creating Layer Convolution2 +I0718 09:46:46.480327 31304 net.cpp:408] Convolution2 <- Pooling1 +I0718 09:46:46.480342 31304 net.cpp:382] Convolution2 -> Convolution2 +I0718 09:46:46.480782 31304 net.cpp:124] Setting up Convolution2 +I0718 09:46:46.480805 31304 net.cpp:131] Top shape: 50 50 8 8 (160000) +I0718 09:46:46.480820 31304 net.cpp:139] Memory required for data: 3677000 +I0718 09:46:46.480839 31304 layer_factory.hpp:77] Creating layer Pooling2 +I0718 09:46:46.480865 31304 net.cpp:86] Creating Layer Pooling2 +I0718 09:46:46.480875 31304 net.cpp:408] Pooling2 <- Convolution2 +I0718 09:46:46.480890 31304 net.cpp:382] Pooling2 -> Pooling2 +I0718 09:46:46.480914 31304 net.cpp:124] Setting up Pooling2 +I0718 09:46:46.480926 31304 net.cpp:131] Top shape: 50 50 4 4 (40000) +I0718 09:46:46.480940 31304 net.cpp:139] Memory required for data: 3837000 +I0718 09:46:46.480949 31304 layer_factory.hpp:77] Creating layer InnerProduct1 +I0718 09:46:46.480965 31304 net.cpp:86] Creating Layer InnerProduct1 +I0718 09:46:46.480974 31304 net.cpp:408] InnerProduct1 <- Pooling2 +I0718 09:46:46.480989 31304 net.cpp:382] InnerProduct1 -> InnerProduct1 +I0718 09:46:46.487133 31304 net.cpp:124] Setting up InnerProduct1 +I0718 09:46:46.487277 31304 net.cpp:131] Top shape: 50 500 (25000) +I0718 09:46:46.487306 31304 net.cpp:139] Memory required for data: 3937000 +I0718 09:46:46.487337 31304 layer_factory.hpp:77] Creating layer ReLU1 +I0718 09:46:46.487366 31304 net.cpp:86] Creating Layer ReLU1 +I0718 09:46:46.487380 31304 net.cpp:408] ReLU1 <- InnerProduct1 +I0718 09:46:46.487399 31304 net.cpp:369] ReLU1 -> InnerProduct1 (in-place) +I0718 09:46:46.487426 31304 net.cpp:124] Setting up ReLU1 +I0718 09:46:46.487437 31304 net.cpp:131] Top shape: 50 500 (25000) +I0718 09:46:46.487448 31304 net.cpp:139] Memory required for data: 4037000 +I0718 09:46:46.487458 31304 layer_factory.hpp:77] Creating layer InnerProduct2 +I0718 09:46:46.487474 31304 net.cpp:86] Creating Layer InnerProduct2 +I0718 09:46:46.487483 31304 net.cpp:408] InnerProduct2 <- InnerProduct1 +I0718 09:46:46.487499 31304 net.cpp:382] InnerProduct2 -> InnerProduct2 +I0718 09:46:46.487627 31304 net.cpp:124] Setting up InnerProduct2 +I0718 09:46:46.487643 31304 net.cpp:131] Top shape: 50 10 (500) +I0718 09:46:46.487655 31304 net.cpp:139] Memory required for data: 4039000 +I0718 09:46:46.487670 31304 layer_factory.hpp:77] Creating layer SoftmaxWithLoss1 +I0718 09:46:46.487694 31304 net.cpp:86] Creating Layer SoftmaxWithLoss1 +I0718 09:46:46.487704 31304 net.cpp:408] SoftmaxWithLoss1 <- InnerProduct2 +I0718 09:46:46.487715 31304 net.cpp:408] SoftmaxWithLoss1 <- DummyData2 +I0718 09:46:46.487731 31304 net.cpp:382] SoftmaxWithLoss1 -> SoftmaxWithLoss1 +I0718 09:46:46.487758 31304 layer_factory.hpp:77] Creating layer SoftmaxWithLoss1 +I0718 09:46:46.487802 31304 net.cpp:124] Setting up SoftmaxWithLoss1 +I0718 09:46:46.487813 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.487825 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.487857 31304 net.cpp:139] Memory required for data: 4039004 +I0718 09:46:46.487867 31304 net.cpp:200] SoftmaxWithLoss1 needs backward computation. +I0718 09:46:46.487879 31304 net.cpp:200] InnerProduct2 needs backward computation. +I0718 09:46:46.487887 31304 net.cpp:200] ReLU1 needs backward computation. +I0718 09:46:46.487896 31304 net.cpp:200] InnerProduct1 needs backward computation. +I0718 09:46:46.487906 31304 net.cpp:200] Pooling2 needs backward computation. +I0718 09:46:46.487916 31304 net.cpp:200] Convolution2 needs backward computation. +I0718 09:46:46.487926 31304 net.cpp:200] Pooling1 needs backward computation. +I0718 09:46:46.487936 31304 net.cpp:200] Convolution1 needs backward computation. +I0718 09:46:46.487946 31304 net.cpp:202] DummyData1 does not need backward computation. +I0718 09:46:46.487955 31304 net.cpp:244] This network produces output SoftmaxWithLoss1 +I0718 09:46:46.487977 31304 net.cpp:257] Network initialization done. ./build/caffe-1.0.0+git20180821.99bd997/python/caffe/test/test_net_spec.py:87: DeprecationWarning: Please use assertRaisesRegex instead. with self.assertRaisesRegexp(TypeError, r): -.I0819 13:44:52.106621 13551 net.cpp:53] Initializing net from parameters: +.I0718 09:46:46.492659 31304 net.cpp:53] Initializing net from parameters: state { phase: TEST level: 0 @@ -17520,32 +17587,32 @@ type: "Silence" bottom: "data2" } -I0819 13:44:52.106763 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.106787 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.106803 13551 net.cpp:382] data -> data -I0819 13:44:52.106832 13551 net.cpp:382] data -> data2 -I0819 13:44:52.106870 13551 net.cpp:124] Setting up data -I0819 13:44:52.106882 13551 net.cpp:131] Top shape: 3 (3) -I0819 13:44:52.106896 13551 net.cpp:131] Top shape: 3 (3) -I0819 13:44:52.106905 13551 net.cpp:139] Memory required for data: 24 -I0819 13:44:52.106914 13551 layer_factory.hpp:77] Creating layer silence_data -I0819 13:44:52.106947 13551 net.cpp:86] Creating Layer silence_data -I0819 13:44:52.106961 13551 net.cpp:408] silence_data <- data -I0819 13:44:52.106974 13551 net.cpp:124] Setting up silence_data -I0819 13:44:52.106982 13551 net.cpp:139] Memory required for data: 24 -I0819 13:44:52.106990 13551 layer_factory.hpp:77] Creating layer silence_data2 -I0819 13:44:52.107002 13551 net.cpp:86] Creating Layer silence_data2 -I0819 13:44:52.107012 13551 net.cpp:408] silence_data2 <- data2 -I0819 13:44:52.107024 13551 net.cpp:124] Setting up silence_data2 -I0819 13:44:52.107033 13551 net.cpp:139] Memory required for data: 24 -I0819 13:44:52.107041 13551 net.cpp:202] silence_data2 does not need backward computation. -I0819 13:44:52.107050 13551 net.cpp:202] silence_data does not need backward computation. -I0819 13:44:52.107061 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.107074 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.108374 13551 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmpv7h6l0ca -I0819 13:44:52.108409 13551 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. -W0819 13:44:52.108418 13551 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. -I0819 13:44:52.108458 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.492820 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.492848 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.492864 31304 net.cpp:382] data -> data +I0718 09:46:46.492894 31304 net.cpp:382] data -> data2 +I0718 09:46:46.492938 31304 net.cpp:124] Setting up data +I0718 09:46:46.492950 31304 net.cpp:131] Top shape: 3 (3) +I0718 09:46:46.492964 31304 net.cpp:131] Top shape: 3 (3) +I0718 09:46:46.492974 31304 net.cpp:139] Memory required for data: 24 +I0718 09:46:46.492983 31304 layer_factory.hpp:77] Creating layer silence_data +I0718 09:46:46.493016 31304 net.cpp:86] Creating Layer silence_data +I0718 09:46:46.493029 31304 net.cpp:408] silence_data <- data +I0718 09:46:46.493044 31304 net.cpp:124] Setting up silence_data +I0718 09:46:46.493052 31304 net.cpp:139] Memory required for data: 24 +I0718 09:46:46.493059 31304 layer_factory.hpp:77] Creating layer silence_data2 +I0718 09:46:46.493072 31304 net.cpp:86] Creating Layer silence_data2 +I0718 09:46:46.493081 31304 net.cpp:408] silence_data2 <- data2 +I0718 09:46:46.493090 31304 net.cpp:124] Setting up silence_data2 +I0718 09:46:46.493098 31304 net.cpp:139] Memory required for data: 24 +I0718 09:46:46.493106 31304 net.cpp:202] silence_data2 does not need backward computation. +I0718 09:46:46.493116 31304 net.cpp:202] silence_data does not need backward computation. +I0718 09:46:46.493125 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.493139 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.494557 31304 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmpej39v5zn +I0718 09:46:46.494601 31304 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. +W0718 09:46:46.494611 31304 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. +I0718 09:46:46.494657 31304 net.cpp:53] Initializing net from parameters: name: "pythonnet" force_backward: true state { @@ -17594,43 +17661,43 @@ layer: "SimpleLayer" } } -I0819 13:44:52.108595 13551 layer_factory.hpp:77] Creating layer input -I0819 13:44:52.108619 13551 net.cpp:86] Creating Layer input -I0819 13:44:52.108634 13551 net.cpp:382] input -> data -I0819 13:44:52.108660 13551 net.cpp:124] Setting up input -I0819 13:44:52.108671 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.108685 13551 net.cpp:139] Memory required for data: 2880 -I0819 13:44:52.108692 13551 layer_factory.hpp:77] Creating layer one -I0819 13:44:52.108785 13551 net.cpp:86] Creating Layer one -I0819 13:44:52.108804 13551 net.cpp:408] one <- data -I0819 13:44:52.108822 13551 net.cpp:382] one -> one -I0819 13:44:52.108943 13551 net.cpp:124] Setting up one -I0819 13:44:52.108963 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.108975 13551 net.cpp:139] Memory required for data: 5760 -I0819 13:44:52.108983 13551 layer_factory.hpp:77] Creating layer two -I0819 13:44:52.109035 13551 net.cpp:86] Creating Layer two -I0819 13:44:52.109048 13551 net.cpp:408] two <- one -I0819 13:44:52.109062 13551 net.cpp:382] two -> two -I0819 13:44:52.109126 13551 net.cpp:124] Setting up two -I0819 13:44:52.109144 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.109158 13551 net.cpp:139] Memory required for data: 8640 -I0819 13:44:52.109166 13551 layer_factory.hpp:77] Creating layer three -I0819 13:44:52.109220 13551 net.cpp:86] Creating Layer three -I0819 13:44:52.109233 13551 net.cpp:408] three <- two -I0819 13:44:52.109248 13551 net.cpp:382] three -> three -I0819 13:44:52.109309 13551 net.cpp:124] Setting up three -I0819 13:44:52.109323 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.109335 13551 net.cpp:139] Memory required for data: 11520 -I0819 13:44:52.109344 13551 net.cpp:202] three does not need backward computation. -I0819 13:44:52.109354 13551 net.cpp:202] two does not need backward computation. -I0819 13:44:52.109362 13551 net.cpp:202] one does not need backward computation. -I0819 13:44:52.109371 13551 net.cpp:202] input does not need backward computation. -I0819 13:44:52.109380 13551 net.cpp:244] This network produces output three -I0819 13:44:52.109393 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.120697 13551 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmpija87cp7 -I0819 13:44:52.120744 13551 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. -W0819 13:44:52.120752 13551 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. -I0819 13:44:52.120803 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.494827 31304 layer_factory.hpp:77] Creating layer input +I0718 09:46:46.494856 31304 net.cpp:86] Creating Layer input +I0718 09:46:46.494872 31304 net.cpp:382] input -> data +I0718 09:46:46.494902 31304 net.cpp:124] Setting up input +I0718 09:46:46.494913 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.494927 31304 net.cpp:139] Memory required for data: 2880 +I0718 09:46:46.494938 31304 layer_factory.hpp:77] Creating layer one +I0718 09:46:46.495033 31304 net.cpp:86] Creating Layer one +I0718 09:46:46.495052 31304 net.cpp:408] one <- data +I0718 09:46:46.495070 31304 net.cpp:382] one -> one +I0718 09:46:46.495324 31304 net.cpp:124] Setting up one +I0718 09:46:46.495352 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.495365 31304 net.cpp:139] Memory required for data: 5760 +I0718 09:46:46.495374 31304 layer_factory.hpp:77] Creating layer two +I0718 09:46:46.495440 31304 net.cpp:86] Creating Layer two +I0718 09:46:46.495455 31304 net.cpp:408] two <- one +I0718 09:46:46.495471 31304 net.cpp:382] two -> two +I0718 09:46:46.495543 31304 net.cpp:124] Setting up two +I0718 09:46:46.495559 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.495573 31304 net.cpp:139] Memory required for data: 8640 +I0718 09:46:46.495581 31304 layer_factory.hpp:77] Creating layer three +I0718 09:46:46.495635 31304 net.cpp:86] Creating Layer three +I0718 09:46:46.495651 31304 net.cpp:408] three <- two +I0718 09:46:46.495666 31304 net.cpp:382] three -> three +I0718 09:46:46.495730 31304 net.cpp:124] Setting up three +I0718 09:46:46.495746 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.495759 31304 net.cpp:139] Memory required for data: 11520 +I0718 09:46:46.495769 31304 net.cpp:202] three does not need backward computation. +I0718 09:46:46.495777 31304 net.cpp:202] two does not need backward computation. +I0718 09:46:46.495788 31304 net.cpp:202] one does not need backward computation. +I0718 09:46:46.495798 31304 net.cpp:202] input does not need backward computation. +I0718 09:46:46.495806 31304 net.cpp:244] This network produces output three +I0718 09:46:46.495821 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.507660 31304 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmp92qal6zr +I0718 09:46:46.507726 31304 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. +W0718 09:46:46.507736 31304 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. +I0718 09:46:46.507795 31304 net.cpp:53] Initializing net from parameters: name: "pythonnet" force_backward: true state { @@ -17679,43 +17746,43 @@ layer: "SimpleLayer" } } -I0819 13:44:52.120990 13551 layer_factory.hpp:77] Creating layer input -I0819 13:44:52.121021 13551 net.cpp:86] Creating Layer input -I0819 13:44:52.121038 13551 net.cpp:382] input -> data -I0819 13:44:52.121075 13551 net.cpp:124] Setting up input -I0819 13:44:52.121088 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.121104 13551 net.cpp:139] Memory required for data: 2880 -I0819 13:44:52.121114 13551 layer_factory.hpp:77] Creating layer one -I0819 13:44:52.121213 13551 net.cpp:86] Creating Layer one -I0819 13:44:52.121233 13551 net.cpp:408] one <- data -I0819 13:44:52.121250 13551 net.cpp:382] one -> one -I0819 13:44:52.121358 13551 net.cpp:124] Setting up one -I0819 13:44:52.121378 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.121392 13551 net.cpp:139] Memory required for data: 5760 -I0819 13:44:52.121402 13551 layer_factory.hpp:77] Creating layer two -I0819 13:44:52.121788 13551 net.cpp:86] Creating Layer two -I0819 13:44:52.121814 13551 net.cpp:408] two <- one -I0819 13:44:52.121829 13551 net.cpp:382] two -> two -I0819 13:44:52.121915 13551 net.cpp:124] Setting up two -I0819 13:44:52.121932 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.121946 13551 net.cpp:139] Memory required for data: 8640 -I0819 13:44:52.121955 13551 layer_factory.hpp:77] Creating layer three -I0819 13:44:52.122017 13551 net.cpp:86] Creating Layer three -I0819 13:44:52.122031 13551 net.cpp:408] three <- two -I0819 13:44:52.122046 13551 net.cpp:382] three -> three -I0819 13:44:52.122114 13551 net.cpp:124] Setting up three -I0819 13:44:52.122130 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.122141 13551 net.cpp:139] Memory required for data: 11520 -I0819 13:44:52.122151 13551 net.cpp:202] three does not need backward computation. -I0819 13:44:52.122160 13551 net.cpp:202] two does not need backward computation. -I0819 13:44:52.122169 13551 net.cpp:202] one does not need backward computation. -I0819 13:44:52.122177 13551 net.cpp:202] input does not need backward computation. -I0819 13:44:52.122185 13551 net.cpp:244] This network produces output three -I0819 13:44:52.122200 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.123010 13551 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmpyoh3xwg_ -I0819 13:44:52.123039 13551 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. -W0819 13:44:52.123047 13551 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. -I0819 13:44:52.123078 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.507994 31304 layer_factory.hpp:77] Creating layer input +I0718 09:46:46.508025 31304 net.cpp:86] Creating Layer input +I0718 09:46:46.508044 31304 net.cpp:382] input -> data +I0718 09:46:46.508082 31304 net.cpp:124] Setting up input +I0718 09:46:46.508095 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.508111 31304 net.cpp:139] Memory required for data: 2880 +I0718 09:46:46.508121 31304 layer_factory.hpp:77] Creating layer one +I0718 09:46:46.508219 31304 net.cpp:86] Creating Layer one +I0718 09:46:46.508237 31304 net.cpp:408] one <- data +I0718 09:46:46.508255 31304 net.cpp:382] one -> one +I0718 09:46:46.508369 31304 net.cpp:124] Setting up one +I0718 09:46:46.508388 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.508399 31304 net.cpp:139] Memory required for data: 5760 +I0718 09:46:46.508407 31304 layer_factory.hpp:77] Creating layer two +I0718 09:46:46.508463 31304 net.cpp:86] Creating Layer two +I0718 09:46:46.508476 31304 net.cpp:408] two <- one +I0718 09:46:46.508489 31304 net.cpp:382] two -> two +I0718 09:46:46.508545 31304 net.cpp:124] Setting up two +I0718 09:46:46.508559 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.508570 31304 net.cpp:139] Memory required for data: 8640 +I0718 09:46:46.508579 31304 layer_factory.hpp:77] Creating layer three +I0718 09:46:46.508615 31304 net.cpp:86] Creating Layer three +I0718 09:46:46.508627 31304 net.cpp:408] three <- two +I0718 09:46:46.508639 31304 net.cpp:382] three -> three +I0718 09:46:46.508689 31304 net.cpp:124] Setting up three +I0718 09:46:46.508702 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.508716 31304 net.cpp:139] Memory required for data: 11520 +I0718 09:46:46.508726 31304 net.cpp:202] three does not need backward computation. +I0718 09:46:46.508738 31304 net.cpp:202] two does not need backward computation. +I0718 09:46:46.508749 31304 net.cpp:202] one does not need backward computation. +I0718 09:46:46.508759 31304 net.cpp:202] input does not need backward computation. +I0718 09:46:46.508769 31304 net.cpp:244] This network produces output three +I0718 09:46:46.508785 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.509658 31304 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmpptxuzysb +I0718 09:46:46.509696 31304 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. +W0718 09:46:46.509706 31304 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. +I0718 09:46:46.509742 31304 net.cpp:53] Initializing net from parameters: name: "pythonnet" force_backward: true state { @@ -17744,20 +17811,20 @@ layer: "ExceptionLayer" } } -I0819 13:44:52.123214 13551 layer_factory.hpp:77] Creating layer input -I0819 13:44:52.123239 13551 net.cpp:86] Creating Layer input -I0819 13:44:52.123253 13551 net.cpp:382] input -> data -I0819 13:44:52.123279 13551 net.cpp:124] Setting up input -I0819 13:44:52.123289 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.123301 13551 net.cpp:139] Memory required for data: 2880 -I0819 13:44:52.123311 13551 layer_factory.hpp:77] Creating layer layer -I0819 13:44:52.123389 13551 net.cpp:86] Creating Layer layer -I0819 13:44:52.123405 13551 net.cpp:408] layer <- data -I0819 13:44:52.123421 13551 net.cpp:382] layer -> top -.I0819 13:44:52.124660 13551 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmp7shbeyw8 -I0819 13:44:52.124691 13551 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. -W0819 13:44:52.124701 13551 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. -I0819 13:44:52.124739 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.509873 31304 layer_factory.hpp:77] Creating layer input +I0718 09:46:46.509903 31304 net.cpp:86] Creating Layer input +I0718 09:46:46.509922 31304 net.cpp:382] input -> data +I0718 09:46:46.509954 31304 net.cpp:124] Setting up input +I0718 09:46:46.509971 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.509987 31304 net.cpp:139] Memory required for data: 2880 +I0718 09:46:46.509999 31304 layer_factory.hpp:77] Creating layer layer +I0718 09:46:46.510089 31304 net.cpp:86] Creating Layer layer +I0718 09:46:46.510109 31304 net.cpp:408] layer <- data +I0718 09:46:46.510128 31304 net.cpp:382] layer -> top +.I0718 09:46:46.511584 31304 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmpyo351_he +I0718 09:46:46.511628 31304 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. +W0718 09:46:46.511638 31304 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. +I0718 09:46:46.511682 31304 net.cpp:53] Initializing net from parameters: name: "pythonnet" force_backward: true state { @@ -17806,43 +17873,43 @@ layer: "SimpleLayer" } } -I0819 13:44:52.124877 13551 layer_factory.hpp:77] Creating layer input -I0819 13:44:52.124902 13551 net.cpp:86] Creating Layer input -I0819 13:44:52.124918 13551 net.cpp:382] input -> data -I0819 13:44:52.124945 13551 net.cpp:124] Setting up input -I0819 13:44:52.124958 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.124972 13551 net.cpp:139] Memory required for data: 2880 -I0819 13:44:52.124984 13551 layer_factory.hpp:77] Creating layer one -I0819 13:44:52.125052 13551 net.cpp:86] Creating Layer one -I0819 13:44:52.125070 13551 net.cpp:408] one <- data -I0819 13:44:52.125087 13551 net.cpp:382] one -> one -I0819 13:44:52.125177 13551 net.cpp:124] Setting up one -I0819 13:44:52.125195 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.125207 13551 net.cpp:139] Memory required for data: 5760 -I0819 13:44:52.125217 13551 layer_factory.hpp:77] Creating layer two -I0819 13:44:52.125272 13551 net.cpp:86] Creating Layer two -I0819 13:44:52.125288 13551 net.cpp:408] two <- one -I0819 13:44:52.125303 13551 net.cpp:382] two -> two -I0819 13:44:52.125368 13551 net.cpp:124] Setting up two -I0819 13:44:52.125383 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.125396 13551 net.cpp:139] Memory required for data: 8640 -I0819 13:44:52.125404 13551 layer_factory.hpp:77] Creating layer three -I0819 13:44:52.125795 13551 net.cpp:86] Creating Layer three -I0819 13:44:52.125825 13551 net.cpp:408] three <- two -I0819 13:44:52.125844 13551 net.cpp:382] three -> three -I0819 13:44:52.125921 13551 net.cpp:124] Setting up three -I0819 13:44:52.125936 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.125948 13551 net.cpp:139] Memory required for data: 11520 -I0819 13:44:52.125958 13551 net.cpp:202] three does not need backward computation. -I0819 13:44:52.125967 13551 net.cpp:202] two does not need backward computation. -I0819 13:44:52.125975 13551 net.cpp:202] one does not need backward computation. -I0819 13:44:52.125984 13551 net.cpp:202] input does not need backward computation. -I0819 13:44:52.125993 13551 net.cpp:244] This network produces output three -I0819 13:44:52.126006 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.137224 13551 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmpbmdtsdqk -I0819 13:44:52.137275 13551 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. -W0819 13:44:52.137285 13551 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. -I0819 13:44:52.137336 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.511837 31304 layer_factory.hpp:77] Creating layer input +I0718 09:46:46.511862 31304 net.cpp:86] Creating Layer input +I0718 09:46:46.511876 31304 net.cpp:382] input -> data +I0718 09:46:46.511904 31304 net.cpp:124] Setting up input +I0718 09:46:46.511915 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.511929 31304 net.cpp:139] Memory required for data: 2880 +I0718 09:46:46.511937 31304 layer_factory.hpp:77] Creating layer one +I0718 09:46:46.512010 31304 net.cpp:86] Creating Layer one +I0718 09:46:46.512027 31304 net.cpp:408] one <- data +I0718 09:46:46.512043 31304 net.cpp:382] one -> one +I0718 09:46:46.512145 31304 net.cpp:124] Setting up one +I0718 09:46:46.512162 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.512176 31304 net.cpp:139] Memory required for data: 5760 +I0718 09:46:46.512184 31304 layer_factory.hpp:77] Creating layer two +I0718 09:46:46.512238 31304 net.cpp:86] Creating Layer two +I0718 09:46:46.512252 31304 net.cpp:408] two <- one +I0718 09:46:46.512267 31304 net.cpp:382] two -> two +I0718 09:46:46.512334 31304 net.cpp:124] Setting up two +I0718 09:46:46.512351 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.512365 31304 net.cpp:139] Memory required for data: 8640 +I0718 09:46:46.512373 31304 layer_factory.hpp:77] Creating layer three +I0718 09:46:46.512426 31304 net.cpp:86] Creating Layer three +I0718 09:46:46.512441 31304 net.cpp:408] three <- two +I0718 09:46:46.512457 31304 net.cpp:382] three -> three +I0718 09:46:46.512524 31304 net.cpp:124] Setting up three +I0718 09:46:46.512540 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.512553 31304 net.cpp:139] Memory required for data: 11520 +I0718 09:46:46.512563 31304 net.cpp:202] three does not need backward computation. +I0718 09:46:46.512573 31304 net.cpp:202] two does not need backward computation. +I0718 09:46:46.512583 31304 net.cpp:202] one does not need backward computation. +I0718 09:46:46.512593 31304 net.cpp:202] input does not need backward computation. +I0718 09:46:46.512601 31304 net.cpp:244] This network produces output three +I0718 09:46:46.512616 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.524626 31304 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmp61t0zybz +I0718 09:46:46.524675 31304 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. +W0718 09:46:46.524684 31304 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. +I0718 09:46:46.524745 31304 net.cpp:53] Initializing net from parameters: name: "pythonnet" force_backward: true state { @@ -17891,43 +17958,43 @@ layer: "SimpleLayer" } } -I0819 13:44:52.137811 13551 layer_factory.hpp:77] Creating layer input -I0819 13:44:52.137841 13551 net.cpp:86] Creating Layer input -I0819 13:44:52.137857 13551 net.cpp:382] input -> data -I0819 13:44:52.137892 13551 net.cpp:124] Setting up input -I0819 13:44:52.137903 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.137917 13551 net.cpp:139] Memory required for data: 2880 -I0819 13:44:52.137928 13551 layer_factory.hpp:77] Creating layer one -I0819 13:44:52.138015 13551 net.cpp:86] Creating Layer one -I0819 13:44:52.138031 13551 net.cpp:408] one <- data -I0819 13:44:52.138048 13551 net.cpp:382] one -> one -I0819 13:44:52.138152 13551 net.cpp:124] Setting up one -I0819 13:44:52.138170 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.138181 13551 net.cpp:139] Memory required for data: 5760 -I0819 13:44:52.138190 13551 layer_factory.hpp:77] Creating layer two -I0819 13:44:52.138242 13551 net.cpp:86] Creating Layer two -I0819 13:44:52.138257 13551 net.cpp:408] two <- one -I0819 13:44:52.138271 13551 net.cpp:382] two -> two -I0819 13:44:52.138336 13551 net.cpp:124] Setting up two -I0819 13:44:52.138351 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.138363 13551 net.cpp:139] Memory required for data: 8640 -I0819 13:44:52.138371 13551 layer_factory.hpp:77] Creating layer three -I0819 13:44:52.138417 13551 net.cpp:86] Creating Layer three -I0819 13:44:52.138432 13551 net.cpp:408] three <- two -I0819 13:44:52.138445 13551 net.cpp:382] three -> three -I0819 13:44:52.138506 13551 net.cpp:124] Setting up three -I0819 13:44:52.138520 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.138532 13551 net.cpp:139] Memory required for data: 11520 -I0819 13:44:52.138542 13551 net.cpp:202] three does not need backward computation. -I0819 13:44:52.138551 13551 net.cpp:202] two does not need backward computation. -I0819 13:44:52.138561 13551 net.cpp:202] one does not need backward computation. -I0819 13:44:52.138569 13551 net.cpp:202] input does not need backward computation. -I0819 13:44:52.138577 13551 net.cpp:244] This network produces output three -I0819 13:44:52.138592 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.139384 13551 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmpc70r4ghg -I0819 13:44:52.139415 13551 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. -W0819 13:44:52.139425 13551 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. -I0819 13:44:52.139453 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.524960 31304 layer_factory.hpp:77] Creating layer input +I0718 09:46:46.524989 31304 net.cpp:86] Creating Layer input +I0718 09:46:46.525007 31304 net.cpp:382] input -> data +I0718 09:46:46.525045 31304 net.cpp:124] Setting up input +I0718 09:46:46.525059 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.525075 31304 net.cpp:139] Memory required for data: 2880 +I0718 09:46:46.525086 31304 layer_factory.hpp:77] Creating layer one +I0718 09:46:46.525183 31304 net.cpp:86] Creating Layer one +I0718 09:46:46.525202 31304 net.cpp:408] one <- data +I0718 09:46:46.525221 31304 net.cpp:382] one -> one +I0718 09:46:46.525337 31304 net.cpp:124] Setting up one +I0718 09:46:46.525358 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.525372 31304 net.cpp:139] Memory required for data: 5760 +I0718 09:46:46.525380 31304 layer_factory.hpp:77] Creating layer two +I0718 09:46:46.525434 31304 net.cpp:86] Creating Layer two +I0718 09:46:46.525449 31304 net.cpp:408] two <- one +I0718 09:46:46.525463 31304 net.cpp:382] two -> two +I0718 09:46:46.525533 31304 net.cpp:124] Setting up two +I0718 09:46:46.525550 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.525564 31304 net.cpp:139] Memory required for data: 8640 +I0718 09:46:46.525573 31304 layer_factory.hpp:77] Creating layer three +I0718 09:46:46.525621 31304 net.cpp:86] Creating Layer three +I0718 09:46:46.525635 31304 net.cpp:408] three <- two +I0718 09:46:46.525651 31304 net.cpp:382] three -> three +I0718 09:46:46.525720 31304 net.cpp:124] Setting up three +I0718 09:46:46.525738 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.525753 31304 net.cpp:139] Memory required for data: 11520 +I0718 09:46:46.525763 31304 net.cpp:202] three does not need backward computation. +I0718 09:46:46.525772 31304 net.cpp:202] two does not need backward computation. +I0718 09:46:46.525782 31304 net.cpp:202] one does not need backward computation. +I0718 09:46:46.525792 31304 net.cpp:202] input does not need backward computation. +I0718 09:46:46.525801 31304 net.cpp:244] This network produces output three +I0718 09:46:46.525817 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.526605 31304 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmp8rfa_jj5 +I0718 09:46:46.526638 31304 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. +W0718 09:46:46.526646 31304 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. +I0718 09:46:46.526679 31304 net.cpp:53] Initializing net from parameters: name: "pythonnet" force_backward: true state { @@ -17956,27 +18023,27 @@ layer: "ParameterLayer" } } -I0819 13:44:52.139568 13551 layer_factory.hpp:77] Creating layer input -I0819 13:44:52.139593 13551 net.cpp:86] Creating Layer input -I0819 13:44:52.139609 13551 net.cpp:382] input -> data -I0819 13:44:52.139636 13551 net.cpp:124] Setting up input -I0819 13:44:52.139647 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.139660 13551 net.cpp:139] Memory required for data: 2880 -I0819 13:44:52.139670 13551 layer_factory.hpp:77] Creating layer layer -I0819 13:44:52.139750 13551 net.cpp:86] Creating Layer layer -I0819 13:44:52.139766 13551 net.cpp:408] layer <- data -I0819 13:44:52.139781 13551 net.cpp:382] layer -> top -I0819 13:44:52.139923 13551 net.cpp:124] Setting up layer -I0819 13:44:52.139940 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.139953 13551 net.cpp:139] Memory required for data: 5760 -I0819 13:44:52.139978 13551 net.cpp:202] layer does not need backward computation. -I0819 13:44:52.139988 13551 net.cpp:202] input does not need backward computation. -I0819 13:44:52.139997 13551 net.cpp:244] This network produces output top -I0819 13:44:52.140008 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.141865 13551 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmpc70r4ghg -I0819 13:44:52.141916 13551 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. -W0819 13:44:52.141927 13551 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. -I0819 13:44:52.141963 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.526803 31304 layer_factory.hpp:77] Creating layer input +I0718 09:46:46.526826 31304 net.cpp:86] Creating Layer input +I0718 09:46:46.526841 31304 net.cpp:382] input -> data +I0718 09:46:46.526867 31304 net.cpp:124] Setting up input +I0718 09:46:46.526880 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.526892 31304 net.cpp:139] Memory required for data: 2880 +I0718 09:46:46.526902 31304 layer_factory.hpp:77] Creating layer layer +I0718 09:46:46.526968 31304 net.cpp:86] Creating Layer layer +I0718 09:46:46.526986 31304 net.cpp:408] layer <- data +I0718 09:46:46.527004 31304 net.cpp:382] layer -> top +I0718 09:46:46.527160 31304 net.cpp:124] Setting up layer +I0718 09:46:46.527181 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.527359 31304 net.cpp:139] Memory required for data: 5760 +I0718 09:46:46.527388 31304 net.cpp:202] layer does not need backward computation. +I0718 09:46:46.527400 31304 net.cpp:202] input does not need backward computation. +I0718 09:46:46.527408 31304 net.cpp:244] This network produces output top +I0718 09:46:46.527421 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.528980 31304 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmp8rfa_jj5 +I0718 09:46:46.529012 31304 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. +W0718 09:46:46.529021 31304 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. +I0718 09:46:46.529052 31304 net.cpp:53] Initializing net from parameters: name: "pythonnet" force_backward: true state { @@ -18005,27 +18072,27 @@ layer: "ParameterLayer" } } -I0819 13:44:52.142099 13551 layer_factory.hpp:77] Creating layer input -I0819 13:44:52.142125 13551 net.cpp:86] Creating Layer input -I0819 13:44:52.142140 13551 net.cpp:382] input -> data -I0819 13:44:52.142169 13551 net.cpp:124] Setting up input -I0819 13:44:52.142181 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.142195 13551 net.cpp:139] Memory required for data: 2880 -I0819 13:44:52.142205 13551 layer_factory.hpp:77] Creating layer layer -I0819 13:44:52.142285 13551 net.cpp:86] Creating Layer layer -I0819 13:44:52.142303 13551 net.cpp:408] layer <- data -I0819 13:44:52.142318 13551 net.cpp:382] layer -> top -I0819 13:44:52.142446 13551 net.cpp:124] Setting up layer -I0819 13:44:52.142462 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.142474 13551 net.cpp:139] Memory required for data: 5760 -I0819 13:44:52.142495 13551 net.cpp:202] layer does not need backward computation. -I0819 13:44:52.142505 13551 net.cpp:202] input does not need backward computation. -I0819 13:44:52.142514 13551 net.cpp:244] This network produces output top -I0819 13:44:52.142529 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.143805 13551 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmpv9oi5n8t -I0819 13:44:52.143839 13551 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. -W0819 13:44:52.143848 13551 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. -I0819 13:44:52.143889 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.529175 31304 layer_factory.hpp:77] Creating layer input +I0718 09:46:46.529201 31304 net.cpp:86] Creating Layer input +I0718 09:46:46.529215 31304 net.cpp:382] input -> data +I0718 09:46:46.529243 31304 net.cpp:124] Setting up input +I0718 09:46:46.529255 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.529268 31304 net.cpp:139] Memory required for data: 2880 +I0718 09:46:46.529279 31304 layer_factory.hpp:77] Creating layer layer +I0718 09:46:46.529352 31304 net.cpp:86] Creating Layer layer +I0718 09:46:46.529371 31304 net.cpp:408] layer <- data +I0718 09:46:46.529388 31304 net.cpp:382] layer -> top +I0718 09:46:46.529510 31304 net.cpp:124] Setting up layer +I0718 09:46:46.529528 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.529541 31304 net.cpp:139] Memory required for data: 5760 +I0718 09:46:46.529562 31304 net.cpp:202] layer does not need backward computation. +I0718 09:46:46.529572 31304 net.cpp:202] input does not need backward computation. +I0718 09:46:46.529582 31304 net.cpp:244] This network produces output top +I0718 09:46:46.529595 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.530804 31304 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmpq0ojaxnc +I0718 09:46:46.530835 31304 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. +W0718 09:46:46.530844 31304 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. +I0718 09:46:46.530884 31304 net.cpp:53] Initializing net from parameters: name: "pythonnet" force_backward: true state { @@ -18074,40 +18141,40 @@ layer: "SimpleLayer" } } -I0819 13:44:52.144042 13551 layer_factory.hpp:77] Creating layer input -I0819 13:44:52.144068 13551 net.cpp:86] Creating Layer input -I0819 13:44:52.144083 13551 net.cpp:382] input -> data -I0819 13:44:52.144111 13551 net.cpp:124] Setting up input -I0819 13:44:52.144124 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.144136 13551 net.cpp:139] Memory required for data: 2880 -I0819 13:44:52.144146 13551 layer_factory.hpp:77] Creating layer one -I0819 13:44:52.144219 13551 net.cpp:86] Creating Layer one -I0819 13:44:52.144238 13551 net.cpp:408] one <- data -I0819 13:44:52.144255 13551 net.cpp:382] one -> one -I0819 13:44:52.144346 13551 net.cpp:124] Setting up one -I0819 13:44:52.144364 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.144378 13551 net.cpp:139] Memory required for data: 5760 -I0819 13:44:52.144387 13551 layer_factory.hpp:77] Creating layer two -I0819 13:44:52.144443 13551 net.cpp:86] Creating Layer two -I0819 13:44:52.144459 13551 net.cpp:408] two <- one -I0819 13:44:52.144474 13551 net.cpp:382] two -> two -I0819 13:44:52.144546 13551 net.cpp:124] Setting up two -I0819 13:44:52.144564 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.144577 13551 net.cpp:139] Memory required for data: 8640 -I0819 13:44:52.144587 13551 layer_factory.hpp:77] Creating layer three -I0819 13:44:52.144639 13551 net.cpp:86] Creating Layer three -I0819 13:44:52.144654 13551 net.cpp:408] three <- two -I0819 13:44:52.144670 13551 net.cpp:382] three -> three -I0819 13:44:52.144739 13551 net.cpp:124] Setting up three -I0819 13:44:52.144755 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.144768 13551 net.cpp:139] Memory required for data: 11520 -I0819 13:44:52.144778 13551 net.cpp:202] three does not need backward computation. -I0819 13:44:52.144788 13551 net.cpp:202] two does not need backward computation. -I0819 13:44:52.144798 13551 net.cpp:202] one does not need backward computation. -I0819 13:44:52.144807 13551 net.cpp:202] input does not need backward computation. -I0819 13:44:52.144816 13551 net.cpp:244] This network produces output three -I0819 13:44:52.144830 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.146075 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.531028 31304 layer_factory.hpp:77] Creating layer input +I0718 09:46:46.531054 31304 net.cpp:86] Creating Layer input +I0718 09:46:46.531069 31304 net.cpp:382] input -> data +I0718 09:46:46.531096 31304 net.cpp:124] Setting up input +I0718 09:46:46.531108 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.531122 31304 net.cpp:139] Memory required for data: 2880 +I0718 09:46:46.531132 31304 layer_factory.hpp:77] Creating layer one +I0718 09:46:46.531316 31304 net.cpp:86] Creating Layer one +I0718 09:46:46.531342 31304 net.cpp:408] one <- data +I0718 09:46:46.531360 31304 net.cpp:382] one -> one +I0718 09:46:46.531456 31304 net.cpp:124] Setting up one +I0718 09:46:46.531474 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.531487 31304 net.cpp:139] Memory required for data: 5760 +I0718 09:46:46.531497 31304 layer_factory.hpp:77] Creating layer two +I0718 09:46:46.531553 31304 net.cpp:86] Creating Layer two +I0718 09:46:46.531569 31304 net.cpp:408] two <- one +I0718 09:46:46.531582 31304 net.cpp:382] two -> two +I0718 09:46:46.531647 31304 net.cpp:124] Setting up two +I0718 09:46:46.531662 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.531675 31304 net.cpp:139] Memory required for data: 8640 +I0718 09:46:46.531683 31304 layer_factory.hpp:77] Creating layer three +I0718 09:46:46.531728 31304 net.cpp:86] Creating Layer three +I0718 09:46:46.531741 31304 net.cpp:408] three <- two +I0718 09:46:46.531754 31304 net.cpp:382] three -> three +I0718 09:46:46.531812 31304 net.cpp:124] Setting up three +I0718 09:46:46.531827 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.531839 31304 net.cpp:139] Memory required for data: 11520 +I0718 09:46:46.531848 31304 net.cpp:202] three does not need backward computation. +I0718 09:46:46.531858 31304 net.cpp:202] two does not need backward computation. +I0718 09:46:46.531867 31304 net.cpp:202] one does not need backward computation. +I0718 09:46:46.531875 31304 net.cpp:202] input does not need backward computation. +I0718 09:46:46.531884 31304 net.cpp:244] This network produces output three +I0718 09:46:46.531898 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.532577 31304 net.cpp:53] Initializing net from parameters: name: "pythonnet" force_backward: true state { @@ -18123,16 +18190,16 @@ layer: "PhaseLayer" } } -I0819 13:44:52.146205 13551 layer_factory.hpp:77] Creating layer layer -I0819 13:44:52.146296 13551 net.cpp:86] Creating Layer layer -I0819 13:44:52.146319 13551 net.cpp:382] layer -> phase -I0819 13:44:52.146410 13551 net.cpp:124] Setting up layer -I0819 13:44:52.146427 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.146443 13551 net.cpp:139] Memory required for data: 4 -I0819 13:44:52.146454 13551 net.cpp:202] layer does not need backward computation. -I0819 13:44:52.146463 13551 net.cpp:244] This network produces output phase -I0819 13:44:52.146476 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.147042 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.532663 31304 layer_factory.hpp:77] Creating layer layer +I0718 09:46:46.532732 31304 net.cpp:86] Creating Layer layer +I0718 09:46:46.532750 31304 net.cpp:382] layer -> phase +I0718 09:46:46.532809 31304 net.cpp:124] Setting up layer +I0718 09:46:46.532824 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.532837 31304 net.cpp:139] Memory required for data: 4 +I0718 09:46:46.532847 31304 net.cpp:202] layer does not need backward computation. +I0718 09:46:46.532856 31304 net.cpp:244] This network produces output phase +I0718 09:46:46.532868 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.533337 31304 net.cpp:53] Initializing net from parameters: name: "pythonnet" force_backward: true state { @@ -18148,19 +18215,19 @@ layer: "PhaseLayer" } } -I0819 13:44:52.147135 13551 layer_factory.hpp:77] Creating layer layer -I0819 13:44:52.147207 13551 net.cpp:86] Creating Layer layer -I0819 13:44:52.147228 13551 net.cpp:382] layer -> phase -I0819 13:44:52.147286 13551 net.cpp:124] Setting up layer -I0819 13:44:52.147303 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.147316 13551 net.cpp:139] Memory required for data: 4 -I0819 13:44:52.147326 13551 net.cpp:202] layer does not need backward computation. -I0819 13:44:52.147336 13551 net.cpp:244] This network produces output phase -I0819 13:44:52.147346 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.148680 13551 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmpytxngp_6 -I0819 13:44:52.148715 13551 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. -W0819 13:44:52.148725 13551 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. -I0819 13:44:52.148766 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.533421 31304 layer_factory.hpp:77] Creating layer layer +I0718 09:46:46.533484 31304 net.cpp:86] Creating Layer layer +I0718 09:46:46.533501 31304 net.cpp:382] layer -> phase +I0718 09:46:46.533555 31304 net.cpp:124] Setting up layer +I0718 09:46:46.533567 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.533579 31304 net.cpp:139] Memory required for data: 4 +I0718 09:46:46.533589 31304 net.cpp:202] layer does not need backward computation. +I0718 09:46:46.533598 31304 net.cpp:244] This network produces output phase +I0718 09:46:46.533610 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.534714 31304 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmp349jr2vi +I0718 09:46:46.534749 31304 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. +W0718 09:46:46.534756 31304 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. +I0718 09:46:46.534798 31304 net.cpp:53] Initializing net from parameters: name: "pythonnet" force_backward: true state { @@ -18209,43 +18276,43 @@ layer: "SimpleLayer" } } -I0819 13:44:52.148910 13551 layer_factory.hpp:77] Creating layer input -I0819 13:44:52.148934 13551 net.cpp:86] Creating Layer input -I0819 13:44:52.148950 13551 net.cpp:382] input -> data -I0819 13:44:52.148979 13551 net.cpp:124] Setting up input -I0819 13:44:52.148991 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.149004 13551 net.cpp:139] Memory required for data: 2880 -I0819 13:44:52.149013 13551 layer_factory.hpp:77] Creating layer one -I0819 13:44:52.149078 13551 net.cpp:86] Creating Layer one -I0819 13:44:52.149096 13551 net.cpp:408] one <- data -I0819 13:44:52.149113 13551 net.cpp:382] one -> one -I0819 13:44:52.149205 13551 net.cpp:124] Setting up one -I0819 13:44:52.149222 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.149235 13551 net.cpp:139] Memory required for data: 5760 -I0819 13:44:52.149245 13551 layer_factory.hpp:77] Creating layer two -I0819 13:44:52.149297 13551 net.cpp:86] Creating Layer two -I0819 13:44:52.149312 13551 net.cpp:408] two <- one -I0819 13:44:52.149327 13551 net.cpp:382] two -> two -I0819 13:44:52.149392 13551 net.cpp:124] Setting up two -I0819 13:44:52.149408 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.149420 13551 net.cpp:139] Memory required for data: 8640 -I0819 13:44:52.149685 13551 layer_factory.hpp:77] Creating layer three -I0819 13:44:52.149768 13551 net.cpp:86] Creating Layer three -I0819 13:44:52.149785 13551 net.cpp:408] three <- two -I0819 13:44:52.149801 13551 net.cpp:382] three -> three -I0819 13:44:52.149873 13551 net.cpp:124] Setting up three -I0819 13:44:52.149886 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.149900 13551 net.cpp:139] Memory required for data: 11520 -I0819 13:44:52.149909 13551 net.cpp:202] three does not need backward computation. -I0819 13:44:52.149919 13551 net.cpp:202] two does not need backward computation. -I0819 13:44:52.149927 13551 net.cpp:202] one does not need backward computation. -I0819 13:44:52.149936 13551 net.cpp:202] input does not need backward computation. -I0819 13:44:52.149945 13551 net.cpp:244] This network produces output three -I0819 13:44:52.149957 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.151787 13551 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmp7apq2mfi -I0819 13:44:52.151821 13551 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. -W0819 13:44:52.151830 13551 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. -I0819 13:44:52.151873 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.534938 31304 layer_factory.hpp:77] Creating layer input +I0718 09:46:46.534961 31304 net.cpp:86] Creating Layer input +I0718 09:46:46.534976 31304 net.cpp:382] input -> data +I0718 09:46:46.535001 31304 net.cpp:124] Setting up input +I0718 09:46:46.535012 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.535024 31304 net.cpp:139] Memory required for data: 2880 +I0718 09:46:46.535033 31304 layer_factory.hpp:77] Creating layer one +I0718 09:46:46.535099 31304 net.cpp:86] Creating Layer one +I0718 09:46:46.535115 31304 net.cpp:408] one <- data +I0718 09:46:46.535130 31304 net.cpp:382] one -> one +I0718 09:46:46.535593 31304 net.cpp:124] Setting up one +I0718 09:46:46.535625 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.535638 31304 net.cpp:139] Memory required for data: 5760 +I0718 09:46:46.535647 31304 layer_factory.hpp:77] Creating layer two +I0718 09:46:46.535710 31304 net.cpp:86] Creating Layer two +I0718 09:46:46.535723 31304 net.cpp:408] two <- one +I0718 09:46:46.535738 31304 net.cpp:382] two -> two +I0718 09:46:46.535809 31304 net.cpp:124] Setting up two +I0718 09:46:46.535823 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.535835 31304 net.cpp:139] Memory required for data: 8640 +I0718 09:46:46.535843 31304 layer_factory.hpp:77] Creating layer three +I0718 09:46:46.535888 31304 net.cpp:86] Creating Layer three +I0718 09:46:46.535902 31304 net.cpp:408] three <- two +I0718 09:46:46.535917 31304 net.cpp:382] three -> three +I0718 09:46:46.535980 31304 net.cpp:124] Setting up three +I0718 09:46:46.535995 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.536006 31304 net.cpp:139] Memory required for data: 11520 +I0718 09:46:46.536015 31304 net.cpp:202] three does not need backward computation. +I0718 09:46:46.536024 31304 net.cpp:202] two does not need backward computation. +I0718 09:46:46.536033 31304 net.cpp:202] one does not need backward computation. +I0718 09:46:46.536042 31304 net.cpp:202] input does not need backward computation. +I0718 09:46:46.536051 31304 net.cpp:244] This network produces output three +I0718 09:46:46.536064 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.537804 31304 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmpud5ce3zd +I0718 09:46:46.537839 31304 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. +W0718 09:46:46.537848 31304 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. +I0718 09:46:46.537886 31304 net.cpp:53] Initializing net from parameters: name: "pythonnet" force_backward: true state { @@ -18286,35 +18353,35 @@ param_str: "2" } } -I0819 13:44:52.152021 13551 layer_factory.hpp:77] Creating layer input -I0819 13:44:52.152045 13551 net.cpp:86] Creating Layer input -I0819 13:44:52.152057 13551 net.cpp:382] input -> data -I0819 13:44:52.152084 13551 net.cpp:124] Setting up input -I0819 13:44:52.152096 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.152107 13551 net.cpp:139] Memory required for data: 2880 -I0819 13:44:52.152117 13551 layer_factory.hpp:77] Creating layer mul10 -I0819 13:44:52.152189 13551 net.cpp:86] Creating Layer mul10 -I0819 13:44:52.152204 13551 net.cpp:408] mul10 <- data -I0819 13:44:52.152217 13551 net.cpp:382] mul10 -> mul10 -I0819 13:44:52.152321 13551 net.cpp:124] Setting up mul10 -I0819 13:44:52.152336 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.152350 13551 net.cpp:139] Memory required for data: 5760 -I0819 13:44:52.152359 13551 layer_factory.hpp:77] Creating layer mul2 -I0819 13:44:52.152406 13551 net.cpp:86] Creating Layer mul2 -I0819 13:44:52.152420 13551 net.cpp:408] mul2 <- mul10 -I0819 13:44:52.152432 13551 net.cpp:382] mul2 -> mul2 -I0819 13:44:52.152490 13551 net.cpp:124] Setting up mul2 -I0819 13:44:52.152504 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.152515 13551 net.cpp:139] Memory required for data: 8640 -I0819 13:44:52.152523 13551 net.cpp:202] mul2 does not need backward computation. -I0819 13:44:52.152532 13551 net.cpp:202] mul10 does not need backward computation. -I0819 13:44:52.152540 13551 net.cpp:202] input does not need backward computation. -I0819 13:44:52.152547 13551 net.cpp:244] This network produces output mul2 -I0819 13:44:52.152559 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.163633 13551 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmpip0r__zt -I0819 13:44:52.163683 13551 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. -W0819 13:44:52.163692 13551 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. -I0819 13:44:52.163740 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.538022 31304 layer_factory.hpp:77] Creating layer input +I0718 09:46:46.538048 31304 net.cpp:86] Creating Layer input +I0718 09:46:46.538060 31304 net.cpp:382] input -> data +I0718 09:46:46.538087 31304 net.cpp:124] Setting up input +I0718 09:46:46.538098 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.538112 31304 net.cpp:139] Memory required for data: 2880 +I0718 09:46:46.538122 31304 layer_factory.hpp:77] Creating layer mul10 +I0718 09:46:46.538192 31304 net.cpp:86] Creating Layer mul10 +I0718 09:46:46.538208 31304 net.cpp:408] mul10 <- data +I0718 09:46:46.538223 31304 net.cpp:382] mul10 -> mul10 +I0718 09:46:46.538316 31304 net.cpp:124] Setting up mul10 +I0718 09:46:46.538332 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.538345 31304 net.cpp:139] Memory required for data: 5760 +I0718 09:46:46.538354 31304 layer_factory.hpp:77] Creating layer mul2 +I0718 09:46:46.538404 31304 net.cpp:86] Creating Layer mul2 +I0718 09:46:46.538419 31304 net.cpp:408] mul2 <- mul10 +I0718 09:46:46.538434 31304 net.cpp:382] mul2 -> mul2 +I0718 09:46:46.538502 31304 net.cpp:124] Setting up mul2 +I0718 09:46:46.538520 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.538532 31304 net.cpp:139] Memory required for data: 8640 +I0718 09:46:46.538543 31304 net.cpp:202] mul2 does not need backward computation. +I0718 09:46:46.538553 31304 net.cpp:202] mul10 does not need backward computation. +I0718 09:46:46.538563 31304 net.cpp:202] input does not need backward computation. +I0718 09:46:46.538573 31304 net.cpp:244] This network produces output mul2 +I0718 09:46:46.538587 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.550012 31304 upgrade_proto.cpp:69] Attempting to upgrade input file specified using deprecated input fields: /tmp/tmpyegeis7u +I0718 09:46:46.550065 31304 upgrade_proto.cpp:72] Successfully upgraded file specified using deprecated input fields. +W0718 09:46:46.550076 31304 upgrade_proto.cpp:74] Note that future Caffe releases will only support input layers and not input fields. +I0718 09:46:46.550134 31304 net.cpp:53] Initializing net from parameters: name: "pythonnet" force_backward: true state { @@ -18355,32 +18422,32 @@ param_str: "2" } } -I0819 13:44:52.163914 13551 layer_factory.hpp:77] Creating layer input -I0819 13:44:52.163947 13551 net.cpp:86] Creating Layer input -I0819 13:44:52.163964 13551 net.cpp:382] input -> data -I0819 13:44:52.164002 13551 net.cpp:124] Setting up input -I0819 13:44:52.164014 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.164031 13551 net.cpp:139] Memory required for data: 2880 -I0819 13:44:52.164041 13551 layer_factory.hpp:77] Creating layer mul10 -I0819 13:44:52.164126 13551 net.cpp:86] Creating Layer mul10 -I0819 13:44:52.164144 13551 net.cpp:408] mul10 <- data -I0819 13:44:52.164160 13551 net.cpp:382] mul10 -> mul10 -I0819 13:44:52.164273 13551 net.cpp:124] Setting up mul10 -I0819 13:44:52.164289 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.164304 13551 net.cpp:139] Memory required for data: 5760 -I0819 13:44:52.164314 13551 layer_factory.hpp:77] Creating layer mul2 -I0819 13:44:52.164372 13551 net.cpp:86] Creating Layer mul2 -I0819 13:44:52.164391 13551 net.cpp:408] mul2 <- mul10 -I0819 13:44:52.164407 13551 net.cpp:382] mul2 -> mul2 -I0819 13:44:52.164482 13551 net.cpp:124] Setting up mul2 -I0819 13:44:52.164499 13551 net.cpp:131] Top shape: 10 9 8 (720) -I0819 13:44:52.164512 13551 net.cpp:139] Memory required for data: 8640 -I0819 13:44:52.164521 13551 net.cpp:202] mul2 does not need backward computation. -I0819 13:44:52.164531 13551 net.cpp:202] mul10 does not need backward computation. -I0819 13:44:52.164541 13551 net.cpp:202] input does not need backward computation. -I0819 13:44:52.164549 13551 net.cpp:244] This network produces output mul2 -I0819 13:44:52.164562 13551 net.cpp:257] Network initialization done. -.I0819 13:44:52.176374 13551 solver.cpp:45] Initializing solver from parameters: +I0718 09:46:46.550331 31304 layer_factory.hpp:77] Creating layer input +I0718 09:46:46.550364 31304 net.cpp:86] Creating Layer input +I0718 09:46:46.550380 31304 net.cpp:382] input -> data +I0718 09:46:46.550417 31304 net.cpp:124] Setting up input +I0718 09:46:46.550432 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.550449 31304 net.cpp:139] Memory required for data: 2880 +I0718 09:46:46.550460 31304 layer_factory.hpp:77] Creating layer mul10 +I0718 09:46:46.550559 31304 net.cpp:86] Creating Layer mul10 +I0718 09:46:46.550580 31304 net.cpp:408] mul10 <- data +I0718 09:46:46.550596 31304 net.cpp:382] mul10 -> mul10 +I0718 09:46:46.550721 31304 net.cpp:124] Setting up mul10 +I0718 09:46:46.550741 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.550755 31304 net.cpp:139] Memory required for data: 5760 +I0718 09:46:46.550763 31304 layer_factory.hpp:77] Creating layer mul2 +I0718 09:46:46.550818 31304 net.cpp:86] Creating Layer mul2 +I0718 09:46:46.550833 31304 net.cpp:408] mul2 <- mul10 +I0718 09:46:46.550849 31304 net.cpp:382] mul2 -> mul2 +I0718 09:46:46.550925 31304 net.cpp:124] Setting up mul2 +I0718 09:46:46.550942 31304 net.cpp:131] Top shape: 10 9 8 (720) +I0718 09:46:46.550956 31304 net.cpp:139] Memory required for data: 8640 +I0718 09:46:46.550964 31304 net.cpp:202] mul2 does not need backward computation. +I0718 09:46:46.550976 31304 net.cpp:202] mul10 does not need backward computation. +I0718 09:46:46.550987 31304 net.cpp:202] input does not need backward computation. +I0718 09:46:46.550995 31304 net.cpp:244] This network produces output mul2 +I0718 09:46:46.551009 31304 net.cpp:257] Network initialization done. +.I0718 09:46:46.562685 31304 solver.cpp:45] Initializing solver from parameters: test_iter: 10 test_interval: 10 base_lr: 0.01 @@ -18392,10 +18459,10 @@ momentum: 0.9 weight_decay: 0.0005 snapshot_prefix: "model" -net: "/tmp/tmp8bkiahy5" +net: "/tmp/tmpbqcpkbu9" snapshot_after_train: false -I0819 13:44:52.176548 13551 solver.cpp:102] Creating training net from net file: /tmp/tmp8bkiahy5 -I0819 13:44:52.176895 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.562845 31304 solver.cpp:102] Creating training net from net file: /tmp/tmpbqcpkbu9 +I0718 09:46:46.563158 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -18473,46 +18540,46 @@ bottom: "label" top: "loss" } -I0819 13:44:52.177160 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.177191 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.177210 13551 net.cpp:382] data -> data -I0819 13:44:52.177245 13551 net.cpp:382] data -> label -I0819 13:44:52.177287 13551 net.cpp:124] Setting up data -I0819 13:44:52.177301 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.177318 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.177330 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.177340 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.177366 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.177379 13551 net.cpp:408] conv <- data -I0819 13:44:52.177398 13551 net.cpp:382] conv -> conv -I0819 13:44:52.177779 13551 net.cpp:124] Setting up conv -I0819 13:44:52.177809 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.177822 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.177850 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.177871 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.177882 13551 net.cpp:408] ip <- conv -I0819 13:44:52.177901 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.178174 13551 net.cpp:124] Setting up ip -I0819 13:44:52.178193 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.178205 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.178225 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.178241 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.178251 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.178261 13551 net.cpp:408] loss <- label -I0819 13:44:52.178273 13551 net.cpp:382] loss -> loss -I0819 13:44:52.178294 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.178332 13551 net.cpp:124] Setting up loss -I0819 13:44:52.178342 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.178352 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.178373 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.178382 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.178391 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.178400 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.178409 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.178418 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.178434 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.178731 13551 solver.cpp:190] Creating test net (#0) specified by net file: /tmp/tmp8bkiahy5 -I0819 13:44:52.178786 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.563725 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.563761 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.563781 31304 net.cpp:382] data -> data +I0718 09:46:46.563814 31304 net.cpp:382] data -> label +I0718 09:46:46.563854 31304 net.cpp:124] Setting up data +I0718 09:46:46.563868 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.563885 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.563896 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.563906 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.563935 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.563947 31304 net.cpp:408] conv <- data +I0718 09:46:46.563967 31304 net.cpp:382] conv -> conv +I0718 09:46:46.564055 31304 net.cpp:124] Setting up conv +I0718 09:46:46.564071 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.564085 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.564114 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.564136 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.564147 31304 net.cpp:408] ip <- conv +I0718 09:46:46.564163 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.564440 31304 net.cpp:124] Setting up ip +I0718 09:46:46.564462 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.564474 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.564493 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.564512 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.564522 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.564533 31304 net.cpp:408] loss <- label +I0718 09:46:46.564546 31304 net.cpp:382] loss -> loss +I0718 09:46:46.564571 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.564610 31304 net.cpp:124] Setting up loss +I0718 09:46:46.564621 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.564635 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.564661 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.564672 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.564682 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.564692 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.564703 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.564713 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.564730 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.565022 31304 solver.cpp:190] Creating test net (#0) specified by net file: /tmp/tmpbqcpkbu9 +I0718 09:46:46.565078 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -18590,46 +18657,46 @@ bottom: "label" top: "loss" } -I0819 13:44:52.179003 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.179028 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.179044 13551 net.cpp:382] data -> data -I0819 13:44:52.179066 13551 net.cpp:382] data -> label -I0819 13:44:52.179095 13551 net.cpp:124] Setting up data -I0819 13:44:52.179106 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.179119 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.179131 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.179139 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.179162 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.179173 13551 net.cpp:408] conv <- data -I0819 13:44:52.179190 13551 net.cpp:382] conv -> conv -I0819 13:44:52.179250 13551 net.cpp:124] Setting up conv -I0819 13:44:52.179263 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.179275 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.179293 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.179309 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.179318 13551 net.cpp:408] ip <- conv -I0819 13:44:52.179332 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.179596 13551 net.cpp:124] Setting up ip -I0819 13:44:52.179613 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.179625 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.179642 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.179657 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.179666 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.179677 13551 net.cpp:408] loss <- label -I0819 13:44:52.179689 13551 net.cpp:382] loss -> loss -I0819 13:44:52.179709 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.179744 13551 net.cpp:124] Setting up loss -I0819 13:44:52.179754 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.179766 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.179785 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.179795 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.179803 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.179812 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.179821 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.179831 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.179845 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.179889 13551 solver.cpp:57] Solver scaffolding done. -I0819 13:44:52.180105 13551 solver.cpp:45] Initializing solver from parameters: +I0718 09:46:46.565304 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.565330 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.565346 31304 net.cpp:382] data -> data +I0718 09:46:46.565367 31304 net.cpp:382] data -> label +I0718 09:46:46.565393 31304 net.cpp:124] Setting up data +I0718 09:46:46.565404 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.565418 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.565429 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.565438 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.565460 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.565472 31304 net.cpp:408] conv <- data +I0718 09:46:46.565488 31304 net.cpp:382] conv -> conv +I0718 09:46:46.565547 31304 net.cpp:124] Setting up conv +I0718 09:46:46.565562 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.565575 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.565595 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.565613 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.565623 31304 net.cpp:408] ip <- conv +I0718 09:46:46.565639 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.565909 31304 net.cpp:124] Setting up ip +I0718 09:46:46.565932 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.565943 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.565963 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.565979 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.565989 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.566000 31304 net.cpp:408] loss <- label +I0718 09:46:46.566013 31304 net.cpp:382] loss -> loss +I0718 09:46:46.566035 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.566071 31304 net.cpp:124] Setting up loss +I0718 09:46:46.566084 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.566097 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.566115 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.566128 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.566138 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.566148 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.566157 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.566166 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.566182 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.566231 31304 solver.cpp:57] Solver scaffolding done. +I0718 09:46:46.566440 31304 solver.cpp:45] Initializing solver from parameters: test_iter: 10 test_interval: 10 base_lr: 0.01 @@ -18641,10 +18708,10 @@ momentum: 0.9 weight_decay: 0.0005 snapshot_prefix: "model" -net: "/tmp/tmp8bkiahy5" +net: "/tmp/tmpbqcpkbu9" snapshot_after_train: false -I0819 13:44:52.180193 13551 solver.cpp:102] Creating training net from net file: /tmp/tmp8bkiahy5 -I0819 13:44:52.180430 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.566531 31304 solver.cpp:102] Creating training net from net file: /tmp/tmpbqcpkbu9 +I0718 09:46:46.566784 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -18722,46 +18789,46 @@ bottom: "label" top: "loss" } -I0819 13:44:52.180627 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.180660 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.180680 13551 net.cpp:382] data -> data -I0819 13:44:52.180706 13551 net.cpp:382] data -> label -I0819 13:44:52.180733 13551 net.cpp:124] Setting up data -I0819 13:44:52.180745 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.180760 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.180771 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.180780 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.180805 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.180816 13551 net.cpp:408] conv <- data -I0819 13:44:52.180835 13551 net.cpp:382] conv -> conv -I0819 13:44:52.180893 13551 net.cpp:124] Setting up conv -I0819 13:44:52.180907 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.180922 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.180943 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.180960 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.180970 13551 net.cpp:408] ip <- conv -I0819 13:44:52.180986 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.181258 13551 net.cpp:124] Setting up ip -I0819 13:44:52.181277 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.181288 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.181306 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.181329 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.181337 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.181346 13551 net.cpp:408] loss <- label -I0819 13:44:52.181358 13551 net.cpp:382] loss -> loss -I0819 13:44:52.181378 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.181413 13551 net.cpp:124] Setting up loss -I0819 13:44:52.181424 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.181885 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.181910 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.181918 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.181928 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.181937 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.181946 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.181953 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.181967 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.182233 13551 solver.cpp:190] Creating test net (#0) specified by net file: /tmp/tmp8bkiahy5 -I0819 13:44:52.182287 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.567004 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.567032 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.567050 31304 net.cpp:382] data -> data +I0718 09:46:46.567073 31304 net.cpp:382] data -> label +I0718 09:46:46.567102 31304 net.cpp:124] Setting up data +I0718 09:46:46.567116 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.567131 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.567144 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.567153 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.567179 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.567698 31304 net.cpp:408] conv <- data +I0718 09:46:46.567741 31304 net.cpp:382] conv -> conv +I0718 09:46:46.567814 31304 net.cpp:124] Setting up conv +I0718 09:46:46.567831 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.567844 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.567867 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.567888 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.567898 31304 net.cpp:408] ip <- conv +I0718 09:46:46.567912 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.568181 31304 net.cpp:124] Setting up ip +I0718 09:46:46.568198 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.568210 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.568229 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.568246 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.568256 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.568267 31304 net.cpp:408] loss <- label +I0718 09:46:46.568281 31304 net.cpp:382] loss -> loss +I0718 09:46:46.568303 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.568339 31304 net.cpp:124] Setting up loss +I0718 09:46:46.568352 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.568364 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.568387 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.568397 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.568408 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.568419 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.568429 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.568439 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.568454 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.568749 31304 solver.cpp:190] Creating test net (#0) specified by net file: /tmp/tmpbqcpkbu9 +I0718 09:46:46.568806 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -18839,47 +18906,47 @@ bottom: "label" top: "loss" } -I0819 13:44:52.182478 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.182503 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.182519 13551 net.cpp:382] data -> data -I0819 13:44:52.182540 13551 net.cpp:382] data -> label -I0819 13:44:52.182566 13551 net.cpp:124] Setting up data -I0819 13:44:52.182576 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.182590 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.182600 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.182607 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.182627 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.182636 13551 net.cpp:408] conv <- data -I0819 13:44:52.182651 13551 net.cpp:382] conv -> conv -I0819 13:44:52.182698 13551 net.cpp:124] Setting up conv -I0819 13:44:52.182708 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.182719 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.182737 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.182752 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.182761 13551 net.cpp:408] ip <- conv -I0819 13:44:52.182773 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.183028 13551 net.cpp:124] Setting up ip -I0819 13:44:52.183043 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.183053 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.183069 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.183089 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.183099 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.183108 13551 net.cpp:408] loss <- label -I0819 13:44:52.183120 13551 net.cpp:382] loss -> loss -I0819 13:44:52.183137 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.183167 13551 net.cpp:124] Setting up loss -I0819 13:44:52.183177 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.183187 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.183203 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.183212 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.183221 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.183229 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.183238 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.183248 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.183260 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.183300 13551 solver.cpp:57] Solver scaffolding done. -I0819 13:44:52.184782 13551 sgd_solver.cpp:112] Iteration 0, lr = 0.01 -.I0819 13:44:52.186933 13551 solver.cpp:45] Initializing solver from parameters: +I0718 09:46:46.569029 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.569056 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.569075 31304 net.cpp:382] data -> data +I0718 09:46:46.569099 31304 net.cpp:382] data -> label +I0718 09:46:46.569128 31304 net.cpp:124] Setting up data +I0718 09:46:46.569139 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.569154 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.569165 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.569175 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.569197 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.569209 31304 net.cpp:408] conv <- data +I0718 09:46:46.569226 31304 net.cpp:382] conv -> conv +I0718 09:46:46.569286 31304 net.cpp:124] Setting up conv +I0718 09:46:46.569301 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.569314 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.569334 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.569353 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.569365 31304 net.cpp:408] ip <- conv +I0718 09:46:46.569380 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.569650 31304 net.cpp:124] Setting up ip +I0718 09:46:46.569670 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.569684 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.569705 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.569732 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.569743 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.569756 31304 net.cpp:408] loss <- label +I0718 09:46:46.569770 31304 net.cpp:382] loss -> loss +I0718 09:46:46.569793 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.569826 31304 net.cpp:124] Setting up loss +I0718 09:46:46.569839 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.569852 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.569872 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.569881 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.569892 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.569902 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.569913 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.569923 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.569938 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.569983 31304 solver.cpp:57] Solver scaffolding done. +I0718 09:46:46.571908 31304 sgd_solver.cpp:112] Iteration 0, lr = 0.01 +.I0718 09:46:46.573747 31304 solver.cpp:45] Initializing solver from parameters: test_iter: 10 test_interval: 10 base_lr: 0.01 @@ -18891,10 +18958,10 @@ momentum: 0.9 weight_decay: 0.0005 snapshot_prefix: "model" -net: "/tmp/tmpex61y9jc" +net: "/tmp/tmp87124kdn" snapshot_after_train: false -I0819 13:44:52.187058 13551 solver.cpp:102] Creating training net from net file: /tmp/tmpex61y9jc -I0819 13:44:52.187337 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.573866 31304 solver.cpp:102] Creating training net from net file: /tmp/tmp87124kdn +I0718 09:46:46.574136 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -18972,46 +19039,46 @@ bottom: "label" top: "loss" } -I0819 13:44:52.187578 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.187608 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.187626 13551 net.cpp:382] data -> data -I0819 13:44:52.187652 13551 net.cpp:382] data -> label -I0819 13:44:52.187686 13551 net.cpp:124] Setting up data -I0819 13:44:52.187697 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.187711 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.187721 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.187731 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.187752 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.187762 13551 net.cpp:408] conv <- data -I0819 13:44:52.187777 13551 net.cpp:382] conv -> conv -I0819 13:44:52.187839 13551 net.cpp:124] Setting up conv -I0819 13:44:52.187850 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.187861 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.187882 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.187901 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.187909 13551 net.cpp:408] ip <- conv -I0819 13:44:52.187922 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.188185 13551 net.cpp:124] Setting up ip -I0819 13:44:52.188205 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.188216 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.188233 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.188251 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.188259 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.188270 13551 net.cpp:408] loss <- label -I0819 13:44:52.188282 13551 net.cpp:382] loss -> loss -I0819 13:44:52.188305 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.188340 13551 net.cpp:124] Setting up loss -I0819 13:44:52.188351 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.188362 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.188380 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.188390 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.188400 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.188408 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.188418 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.188426 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.188441 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.188689 13551 solver.cpp:190] Creating test net (#0) specified by net file: /tmp/tmpex61y9jc -I0819 13:44:52.188737 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.574354 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.574379 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.574396 31304 net.cpp:382] data -> data +I0718 09:46:46.574422 31304 net.cpp:382] data -> label +I0718 09:46:46.574455 31304 net.cpp:124] Setting up data +I0718 09:46:46.574465 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.574481 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.574491 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.574501 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.574522 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.574532 31304 net.cpp:408] conv <- data +I0718 09:46:46.574548 31304 net.cpp:382] conv -> conv +I0718 09:46:46.574607 31304 net.cpp:124] Setting up conv +I0718 09:46:46.574620 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.574630 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.574651 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.574667 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.574676 31304 net.cpp:408] ip <- conv +I0718 09:46:46.574690 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.574965 31304 net.cpp:124] Setting up ip +I0718 09:46:46.574987 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.574999 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.575019 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.575038 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.575049 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.575062 31304 net.cpp:408] loss <- label +I0718 09:46:46.575079 31304 net.cpp:382] loss -> loss +I0718 09:46:46.575107 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.575147 31304 net.cpp:124] Setting up loss +I0718 09:46:46.575161 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.575174 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.575528 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.575546 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.575560 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.575570 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.575580 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.575592 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.575611 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.575927 31304 solver.cpp:190] Creating test net (#0) specified by net file: /tmp/tmp87124kdn +I0718 09:46:46.575985 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -19089,46 +19156,46 @@ bottom: "label" top: "loss" } -I0819 13:44:52.188948 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.188977 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.188992 13551 net.cpp:382] data -> data -I0819 13:44:52.189014 13551 net.cpp:382] data -> label -I0819 13:44:52.189042 13551 net.cpp:124] Setting up data -I0819 13:44:52.189054 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.189066 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.189076 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.189085 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.189107 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.189117 13551 net.cpp:408] conv <- data -I0819 13:44:52.189131 13551 net.cpp:382] conv -> conv -I0819 13:44:52.189183 13551 net.cpp:124] Setting up conv -I0819 13:44:52.189195 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.189206 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.189225 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.189240 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.189249 13551 net.cpp:408] ip <- conv -I0819 13:44:52.189260 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.189853 13551 net.cpp:124] Setting up ip -I0819 13:44:52.189883 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.189896 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.189916 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.189931 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.189941 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.189952 13551 net.cpp:408] loss <- label -I0819 13:44:52.189965 13551 net.cpp:382] loss -> loss -I0819 13:44:52.189988 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.190021 13551 net.cpp:124] Setting up loss -I0819 13:44:52.190034 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.190047 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.190064 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.190073 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.190083 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.190091 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.190101 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.190109 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.190124 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.190166 13551 solver.cpp:57] Solver scaffolding done. -I0819 13:44:52.190361 13551 solver.cpp:45] Initializing solver from parameters: +I0718 09:46:46.576205 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.576231 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.576247 31304 net.cpp:382] data -> data +I0718 09:46:46.576273 31304 net.cpp:382] data -> label +I0718 09:46:46.576303 31304 net.cpp:124] Setting up data +I0718 09:46:46.576313 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.576328 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.576339 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.576346 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.576367 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.576377 31304 net.cpp:408] conv <- data +I0718 09:46:46.576395 31304 net.cpp:382] conv -> conv +I0718 09:46:46.576452 31304 net.cpp:124] Setting up conv +I0718 09:46:46.576464 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.576476 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.576498 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.576514 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.576524 31304 net.cpp:408] ip <- conv +I0718 09:46:46.576535 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.576819 31304 net.cpp:124] Setting up ip +I0718 09:46:46.576836 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.576848 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.576865 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.576880 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.576889 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.576900 31304 net.cpp:408] loss <- label +I0718 09:46:46.576913 31304 net.cpp:382] loss -> loss +I0718 09:46:46.576933 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.576965 31304 net.cpp:124] Setting up loss +I0718 09:46:46.576975 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.576987 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.577005 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.577014 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.577024 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.577033 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.577041 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.577052 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.577066 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.577113 31304 solver.cpp:57] Solver scaffolding done. +I0718 09:46:46.577324 31304 solver.cpp:45] Initializing solver from parameters: test_iter: 10 test_interval: 10 base_lr: 0.01 @@ -19140,10 +19207,10 @@ momentum: 0.9 weight_decay: 0.0005 snapshot_prefix: "model" -net: "/tmp/tmpex61y9jc" +net: "/tmp/tmp87124kdn" snapshot_after_train: false -I0819 13:44:52.190449 13551 solver.cpp:102] Creating training net from net file: /tmp/tmpex61y9jc -I0819 13:44:52.190683 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.577412 31304 solver.cpp:102] Creating training net from net file: /tmp/tmp87124kdn +I0718 09:46:46.577644 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -19221,46 +19288,46 @@ bottom: "label" top: "loss" } -I0819 13:44:52.190882 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.190908 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.190923 13551 net.cpp:382] data -> data -I0819 13:44:52.190946 13551 net.cpp:382] data -> label -I0819 13:44:52.190979 13551 net.cpp:124] Setting up data -I0819 13:44:52.190990 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.191005 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.191016 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.191026 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.191048 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.191059 13551 net.cpp:408] conv <- data -I0819 13:44:52.191076 13551 net.cpp:382] conv -> conv -I0819 13:44:52.191133 13551 net.cpp:124] Setting up conv -I0819 13:44:52.191145 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.191157 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.191177 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.191195 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.191205 13551 net.cpp:408] ip <- conv -I0819 13:44:52.191220 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.191485 13551 net.cpp:124] Setting up ip -I0819 13:44:52.191506 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.191519 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.191536 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.191551 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.191561 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.191571 13551 net.cpp:408] loss <- label -I0819 13:44:52.191586 13551 net.cpp:382] loss -> loss -I0819 13:44:52.191605 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.191635 13551 net.cpp:124] Setting up loss -I0819 13:44:52.191645 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.191656 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.191673 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.191681 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.191690 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.191699 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.191707 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.191715 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.191728 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.191978 13551 solver.cpp:190] Creating test net (#0) specified by net file: /tmp/tmpex61y9jc -I0819 13:44:52.192031 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.577839 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.577863 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.577879 31304 net.cpp:382] data -> data +I0718 09:46:46.577900 31304 net.cpp:382] data -> label +I0718 09:46:46.577926 31304 net.cpp:124] Setting up data +I0718 09:46:46.577936 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.577950 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.577960 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.577968 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.577989 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.577999 31304 net.cpp:408] conv <- data +I0718 09:46:46.578014 31304 net.cpp:382] conv -> conv +I0718 09:46:46.578068 31304 net.cpp:124] Setting up conv +I0718 09:46:46.578080 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.578091 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.578109 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.578125 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.578133 31304 net.cpp:408] ip <- conv +I0718 09:46:46.578148 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.578415 31304 net.cpp:124] Setting up ip +I0718 09:46:46.578434 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.578446 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.578465 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.578481 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.578490 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.578500 31304 net.cpp:408] loss <- label +I0718 09:46:46.578513 31304 net.cpp:382] loss -> loss +I0718 09:46:46.578536 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.578567 31304 net.cpp:124] Setting up loss +I0718 09:46:46.578577 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.578586 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.578605 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.578614 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.578624 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.578632 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.578641 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.578651 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.578665 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.578925 31304 solver.cpp:190] Creating test net (#0) specified by net file: /tmp/tmp87124kdn +I0718 09:46:46.578979 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -19338,46 +19405,46 @@ bottom: "label" top: "loss" } -I0819 13:44:52.192229 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.192252 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.192267 13551 net.cpp:382] data -> data -I0819 13:44:52.192289 13551 net.cpp:382] data -> label -I0819 13:44:52.192314 13551 net.cpp:124] Setting up data -I0819 13:44:52.192325 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.192338 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.192349 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.192358 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.192378 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.192389 13551 net.cpp:408] conv <- data -I0819 13:44:52.192404 13551 net.cpp:382] conv -> conv -I0819 13:44:52.192461 13551 net.cpp:124] Setting up conv -I0819 13:44:52.192476 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.192487 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.192507 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.192526 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.192538 13551 net.cpp:408] ip <- conv -I0819 13:44:52.192553 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.192812 13551 net.cpp:124] Setting up ip -I0819 13:44:52.192827 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.192838 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.192855 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.192869 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.192878 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.192888 13551 net.cpp:408] loss <- label -I0819 13:44:52.192903 13551 net.cpp:382] loss -> loss -I0819 13:44:52.192921 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.192948 13551 net.cpp:124] Setting up loss -I0819 13:44:52.192957 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.192966 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.192983 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.192991 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.193001 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.193008 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.193017 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.193025 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.193038 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.193078 13551 solver.cpp:57] Solver scaffolding done. -.I0819 13:44:52.198762 13551 solver.cpp:45] Initializing solver from parameters: +I0718 09:46:46.579185 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.579720 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.579738 31304 net.cpp:382] data -> data +I0718 09:46:46.579766 31304 net.cpp:382] data -> label +I0718 09:46:46.579797 31304 net.cpp:124] Setting up data +I0718 09:46:46.579809 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.579825 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.579836 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.579846 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.579874 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.579885 31304 net.cpp:408] conv <- data +I0718 09:46:46.579905 31304 net.cpp:382] conv -> conv +I0718 09:46:46.579970 31304 net.cpp:124] Setting up conv +I0718 09:46:46.579985 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.579999 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.580021 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.580041 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.580052 31304 net.cpp:408] ip <- conv +I0718 09:46:46.580067 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.580338 31304 net.cpp:124] Setting up ip +I0718 09:46:46.580360 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.580372 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.580394 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.580410 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.580420 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.580430 31304 net.cpp:408] loss <- label +I0718 09:46:46.580444 31304 net.cpp:382] loss -> loss +I0718 09:46:46.580464 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.580499 31304 net.cpp:124] Setting up loss +I0718 09:46:46.580510 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.580521 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.580545 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.580555 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.580564 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.580574 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.580583 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.580591 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.580606 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.580657 31304 solver.cpp:57] Solver scaffolding done. +.I0718 09:46:46.586395 31304 solver.cpp:45] Initializing solver from parameters: test_iter: 10 test_interval: 10 base_lr: 0.01 @@ -19389,10 +19456,10 @@ momentum: 0.9 weight_decay: 0.0005 snapshot_prefix: "model" -net: "/tmp/tmp5c78guaw" +net: "/tmp/tmpua611aem" snapshot_after_train: false -I0819 13:44:52.198920 13551 solver.cpp:102] Creating training net from net file: /tmp/tmp5c78guaw -I0819 13:44:52.199208 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.586546 31304 solver.cpp:102] Creating training net from net file: /tmp/tmpua611aem +I0718 09:46:46.586864 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -19470,46 +19537,46 @@ bottom: "label" top: "loss" } -I0819 13:44:52.199437 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.199467 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.199486 13551 net.cpp:382] data -> data -I0819 13:44:52.199519 13551 net.cpp:382] data -> label -I0819 13:44:52.199556 13551 net.cpp:124] Setting up data -I0819 13:44:52.199570 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.199586 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.199596 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.199606 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.199630 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.199641 13551 net.cpp:408] conv <- data -I0819 13:44:52.199657 13551 net.cpp:382] conv -> conv -I0819 13:44:52.199721 13551 net.cpp:124] Setting up conv -I0819 13:44:52.199735 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.199748 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.199770 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.199790 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.199800 13551 net.cpp:408] ip <- conv -I0819 13:44:52.199815 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.200088 13551 net.cpp:124] Setting up ip -I0819 13:44:52.200109 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.200119 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.200137 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.200155 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.200162 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.200172 13551 net.cpp:408] loss <- label -I0819 13:44:52.200184 13551 net.cpp:382] loss -> loss -I0819 13:44:52.200206 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.200240 13551 net.cpp:124] Setting up loss -I0819 13:44:52.200249 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.200259 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.200278 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.200286 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.200296 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.200305 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.200314 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.200322 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.200337 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.200613 13551 solver.cpp:190] Creating test net (#0) specified by net file: /tmp/tmp5c78guaw -I0819 13:44:52.200668 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.587110 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.587138 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.587154 31304 net.cpp:382] data -> data +I0718 09:46:46.587184 31304 net.cpp:382] data -> label +I0718 09:46:46.587481 31304 net.cpp:124] Setting up data +I0718 09:46:46.587497 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.587512 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.587522 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.587532 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.587558 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.587569 31304 net.cpp:408] conv <- data +I0718 09:46:46.587587 31304 net.cpp:382] conv -> conv +I0718 09:46:46.587658 31304 net.cpp:124] Setting up conv +I0718 09:46:46.587672 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.587684 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.587708 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.587728 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.587736 31304 net.cpp:408] ip <- conv +I0718 09:46:46.587751 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.588018 31304 net.cpp:124] Setting up ip +I0718 09:46:46.588037 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.588048 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.588066 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.588083 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.588093 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.588104 31304 net.cpp:408] loss <- label +I0718 09:46:46.588116 31304 net.cpp:382] loss -> loss +I0718 09:46:46.588137 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.588172 31304 net.cpp:124] Setting up loss +I0718 09:46:46.588184 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.588194 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.588214 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.588223 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.588232 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.588243 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.588251 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.588260 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.588275 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.588564 31304 solver.cpp:190] Creating test net (#0) specified by net file: /tmp/tmpua611aem +I0718 09:46:46.588618 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -19587,46 +19654,46 @@ bottom: "label" top: "loss" } -I0819 13:44:52.200875 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.200899 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.200914 13551 net.cpp:382] data -> data -I0819 13:44:52.200937 13551 net.cpp:382] data -> label -I0819 13:44:52.200960 13551 net.cpp:124] Setting up data -I0819 13:44:52.200970 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.200984 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.200992 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.201000 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.201020 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.201030 13551 net.cpp:408] conv <- data -I0819 13:44:52.201045 13551 net.cpp:382] conv -> conv -I0819 13:44:52.201094 13551 net.cpp:124] Setting up conv -I0819 13:44:52.201105 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.201117 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.201134 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.201150 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.201160 13551 net.cpp:408] ip <- conv -I0819 13:44:52.201174 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.201783 13551 net.cpp:124] Setting up ip -I0819 13:44:52.201815 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.201829 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.201851 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.201871 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.201881 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.201892 13551 net.cpp:408] loss <- label -I0819 13:44:52.201906 13551 net.cpp:382] loss -> loss -I0819 13:44:52.201927 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.201961 13551 net.cpp:124] Setting up loss -I0819 13:44:52.201972 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.201982 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.202000 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.202010 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.202019 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.202028 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.202039 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.202049 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.202064 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.202109 13551 solver.cpp:57] Solver scaffolding done. -I0819 13:44:52.202308 13551 solver.cpp:45] Initializing solver from parameters: +I0718 09:46:46.588829 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.588853 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.588869 31304 net.cpp:382] data -> data +I0718 09:46:46.588892 31304 net.cpp:382] data -> label +I0718 09:46:46.588918 31304 net.cpp:124] Setting up data +I0718 09:46:46.588928 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.588943 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.588953 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.588961 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.588981 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.588994 31304 net.cpp:408] conv <- data +I0718 09:46:46.589011 31304 net.cpp:382] conv -> conv +I0718 09:46:46.589064 31304 net.cpp:124] Setting up conv +I0718 09:46:46.589076 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.589087 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.589105 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.589123 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.589133 31304 net.cpp:408] ip <- conv +I0718 09:46:46.589146 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.589409 31304 net.cpp:124] Setting up ip +I0718 09:46:46.589428 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.589440 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.589458 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.589473 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.589483 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.589493 31304 net.cpp:408] loss <- label +I0718 09:46:46.589505 31304 net.cpp:382] loss -> loss +I0718 09:46:46.589525 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.589553 31304 net.cpp:124] Setting up loss +I0718 09:46:46.589562 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.589571 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.589588 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.589596 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.589605 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.589613 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.589622 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.589630 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.589644 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.589684 31304 solver.cpp:57] Solver scaffolding done. +I0718 09:46:46.589872 31304 solver.cpp:45] Initializing solver from parameters: test_iter: 10 test_interval: 10 base_lr: 0.01 @@ -19638,10 +19705,10 @@ momentum: 0.9 weight_decay: 0.0005 snapshot_prefix: "model" -net: "/tmp/tmp5c78guaw" +net: "/tmp/tmpua611aem" snapshot_after_train: false -I0819 13:44:52.202404 13551 solver.cpp:102] Creating training net from net file: /tmp/tmp5c78guaw -I0819 13:44:52.202641 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.589954 31304 solver.cpp:102] Creating training net from net file: /tmp/tmpua611aem +I0718 09:46:46.590176 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -19719,46 +19786,46 @@ bottom: "label" top: "loss" } -I0819 13:44:52.202865 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.202893 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.202909 13551 net.cpp:382] data -> data -I0819 13:44:52.202932 13551 net.cpp:382] data -> label -I0819 13:44:52.202963 13551 net.cpp:124] Setting up data -I0819 13:44:52.202975 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.202989 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.203001 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.203009 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.203032 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.203042 13551 net.cpp:408] conv <- data -I0819 13:44:52.203060 13551 net.cpp:382] conv -> conv -I0819 13:44:52.203125 13551 net.cpp:124] Setting up conv -I0819 13:44:52.203141 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.203155 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.203176 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.203195 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.203205 13551 net.cpp:408] ip <- conv -I0819 13:44:52.203220 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.203490 13551 net.cpp:124] Setting up ip -I0819 13:44:52.203511 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.203523 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.203541 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.203557 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.203565 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.203577 13551 net.cpp:408] loss <- label -I0819 13:44:52.203588 13551 net.cpp:382] loss -> loss -I0819 13:44:52.203608 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.203637 13551 net.cpp:124] Setting up loss -I0819 13:44:52.203646 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.203655 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.203673 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.203683 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.203692 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.203702 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.203712 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.203722 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.203735 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.203995 13551 solver.cpp:190] Creating test net (#0) specified by net file: /tmp/tmp5c78guaw -I0819 13:44:52.204047 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.590363 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.590386 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.590401 31304 net.cpp:382] data -> data +I0718 09:46:46.590422 31304 net.cpp:382] data -> label +I0718 09:46:46.590448 31304 net.cpp:124] Setting up data +I0718 09:46:46.590457 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.590471 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.590481 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.590488 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.590507 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.590518 31304 net.cpp:408] conv <- data +I0718 09:46:46.590533 31304 net.cpp:382] conv -> conv +I0718 09:46:46.590584 31304 net.cpp:124] Setting up conv +I0718 09:46:46.590595 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.590605 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.590622 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.590637 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.590646 31304 net.cpp:408] ip <- conv +I0718 09:46:46.590659 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.590917 31304 net.cpp:124] Setting up ip +I0718 09:46:46.590934 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.590945 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.590965 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.590979 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.590988 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.590999 31304 net.cpp:408] loss <- label +I0718 09:46:46.591012 31304 net.cpp:382] loss -> loss +I0718 09:46:46.591032 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.591060 31304 net.cpp:124] Setting up loss +I0718 09:46:46.591070 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.591081 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.591099 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.591109 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.591117 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.591126 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.591136 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.591145 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.591159 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.591869 31304 solver.cpp:190] Creating test net (#0) specified by net file: /tmp/tmpua611aem +I0718 09:46:46.591936 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -19836,48 +19903,48 @@ bottom: "label" top: "loss" } -I0819 13:44:52.204258 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.204283 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.204298 13551 net.cpp:382] data -> data -I0819 13:44:52.204320 13551 net.cpp:382] data -> label -I0819 13:44:52.204345 13551 net.cpp:124] Setting up data -I0819 13:44:52.204356 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.204368 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.204378 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.204387 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.204406 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.204416 13551 net.cpp:408] conv <- data -I0819 13:44:52.204433 13551 net.cpp:382] conv -> conv -I0819 13:44:52.204484 13551 net.cpp:124] Setting up conv -I0819 13:44:52.204499 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.204511 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.204531 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.204550 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.204558 13551 net.cpp:408] ip <- conv -I0819 13:44:52.204571 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.204828 13551 net.cpp:124] Setting up ip -I0819 13:44:52.204843 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.204855 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.204872 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.204886 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.204895 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.204905 13551 net.cpp:408] loss <- label -I0819 13:44:52.204921 13551 net.cpp:382] loss -> loss -I0819 13:44:52.204939 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.204968 13551 net.cpp:124] Setting up loss -I0819 13:44:52.204979 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.204989 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.205008 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.205015 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.205025 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.205034 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.205042 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.205051 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.205065 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.205109 13551 solver.cpp:57] Solver scaffolding done. -I0819 13:44:52.206948 13551 solver.cpp:464] Snapshotting to binary proto file model_iter_0.caffemodel -I0819 13:44:52.207487 13551 sgd_solver.cpp:284] Snapshotting solver state to binary proto file model_iter_0.solverstate -.I0819 13:44:52.209326 13551 solver.cpp:45] Initializing solver from parameters: +I0718 09:46:46.592149 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.592171 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.592186 31304 net.cpp:382] data -> data +I0718 09:46:46.592208 31304 net.cpp:382] data -> label +I0718 09:46:46.592234 31304 net.cpp:124] Setting up data +I0718 09:46:46.592245 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.592259 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.592269 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.592278 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.592300 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.592310 31304 net.cpp:408] conv <- data +I0718 09:46:46.592325 31304 net.cpp:382] conv -> conv +I0718 09:46:46.592379 31304 net.cpp:124] Setting up conv +I0718 09:46:46.592392 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.592403 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.592424 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.592447 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.592458 31304 net.cpp:408] ip <- conv +I0718 09:46:46.592474 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.592751 31304 net.cpp:124] Setting up ip +I0718 09:46:46.592768 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.592782 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.592799 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.592816 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.592826 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.592837 31304 net.cpp:408] loss <- label +I0718 09:46:46.592850 31304 net.cpp:382] loss -> loss +I0718 09:46:46.592871 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.592905 31304 net.cpp:124] Setting up loss +I0718 09:46:46.592917 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.592927 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.592945 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.592954 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.592964 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.592973 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.592983 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.592990 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.593006 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.593053 31304 solver.cpp:57] Solver scaffolding done. +I0718 09:46:46.594444 31304 solver.cpp:464] Snapshotting to binary proto file model_iter_0.caffemodel +I0718 09:46:46.594930 31304 sgd_solver.cpp:284] Snapshotting solver state to binary proto file model_iter_0.solverstate +.I0718 09:46:46.597172 31304 solver.cpp:45] Initializing solver from parameters: test_iter: 10 test_interval: 10 base_lr: 0.01 @@ -19889,10 +19956,10 @@ momentum: 0.9 weight_decay: 0.0005 snapshot_prefix: "model" -net: "/tmp/tmpk6qfc9et" +net: "/tmp/tmp33qhmbtj" snapshot_after_train: false -I0819 13:44:52.209707 13551 solver.cpp:102] Creating training net from net file: /tmp/tmpk6qfc9et -I0819 13:44:52.210034 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.597301 31304 solver.cpp:102] Creating training net from net file: /tmp/tmp33qhmbtj +I0718 09:46:46.597579 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -19970,46 +20037,46 @@ bottom: "label" top: "loss" } -I0819 13:44:52.210279 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.210306 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.210323 13551 net.cpp:382] data -> data -I0819 13:44:52.210351 13551 net.cpp:382] data -> label -I0819 13:44:52.210386 13551 net.cpp:124] Setting up data -I0819 13:44:52.210397 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.210410 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.210420 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.210429 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.210453 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.210464 13551 net.cpp:408] conv <- data -I0819 13:44:52.210479 13551 net.cpp:382] conv -> conv -I0819 13:44:52.210544 13551 net.cpp:124] Setting up conv -I0819 13:44:52.210556 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.210567 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.210590 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.210608 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.210618 13551 net.cpp:408] ip <- conv -I0819 13:44:52.210633 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.210916 13551 net.cpp:124] Setting up ip -I0819 13:44:52.210933 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.210945 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.210963 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.210980 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.210990 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.211000 13551 net.cpp:408] loss <- label -I0819 13:44:52.211014 13551 net.cpp:382] loss -> loss -I0819 13:44:52.211035 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.211071 13551 net.cpp:124] Setting up loss -I0819 13:44:52.211079 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.211091 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.211109 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.211118 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.211128 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.211138 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.211148 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.211156 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.211172 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.211448 13551 solver.cpp:190] Creating test net (#0) specified by net file: /tmp/tmpk6qfc9et -I0819 13:44:52.211501 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.597824 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.597854 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.597873 31304 net.cpp:382] data -> data +I0718 09:46:46.597905 31304 net.cpp:382] data -> label +I0718 09:46:46.597942 31304 net.cpp:124] Setting up data +I0718 09:46:46.597959 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.597975 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.597985 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.597995 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.598021 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.598034 31304 net.cpp:408] conv <- data +I0718 09:46:46.598054 31304 net.cpp:382] conv -> conv +I0718 09:46:46.598129 31304 net.cpp:124] Setting up conv +I0718 09:46:46.598145 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.598158 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.598181 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.598201 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.598210 31304 net.cpp:408] ip <- conv +I0718 09:46:46.598227 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.598497 31304 net.cpp:124] Setting up ip +I0718 09:46:46.598517 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.598529 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.598547 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.598563 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.598574 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.598587 31304 net.cpp:408] loss <- label +I0718 09:46:46.598601 31304 net.cpp:382] loss -> loss +I0718 09:46:46.598624 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.598664 31304 net.cpp:124] Setting up loss +I0718 09:46:46.598676 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.598687 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.598706 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.598716 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.598727 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.598737 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.598745 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.598754 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.598770 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.599030 31304 solver.cpp:190] Creating test net (#0) specified by net file: /tmp/tmp33qhmbtj +I0718 09:46:46.599084 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -20087,46 +20154,46 @@ bottom: "label" top: "loss" } -I0819 13:44:52.211715 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.211741 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.211757 13551 net.cpp:382] data -> data -I0819 13:44:52.211781 13551 net.cpp:382] data -> label -I0819 13:44:52.211810 13551 net.cpp:124] Setting up data -I0819 13:44:52.211822 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.211835 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.211846 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.211855 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.211879 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.211890 13551 net.cpp:408] conv <- data -I0819 13:44:52.211907 13551 net.cpp:382] conv -> conv -I0819 13:44:52.211964 13551 net.cpp:124] Setting up conv -I0819 13:44:52.211979 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.211992 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.212013 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.212031 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.212041 13551 net.cpp:408] ip <- conv -I0819 13:44:52.212057 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.212332 13551 net.cpp:124] Setting up ip -I0819 13:44:52.212353 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.212365 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.212384 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.212400 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.212409 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.212420 13551 net.cpp:408] loss <- label -I0819 13:44:52.212435 13551 net.cpp:382] loss -> loss -I0819 13:44:52.212455 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.212491 13551 net.cpp:124] Setting up loss -I0819 13:44:52.212502 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.212514 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.212534 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.212544 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.212554 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.212563 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.212574 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.212582 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.212599 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.212648 13551 solver.cpp:57] Solver scaffolding done. -I0819 13:44:52.212863 13551 solver.cpp:45] Initializing solver from parameters: +I0718 09:46:46.599651 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.599682 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.599700 31304 net.cpp:382] data -> data +I0718 09:46:46.599725 31304 net.cpp:382] data -> label +I0718 09:46:46.599754 31304 net.cpp:124] Setting up data +I0718 09:46:46.599766 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.599781 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.599792 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.599802 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.599826 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.599838 31304 net.cpp:408] conv <- data +I0718 09:46:46.599854 31304 net.cpp:382] conv -> conv +I0718 09:46:46.599910 31304 net.cpp:124] Setting up conv +I0718 09:46:46.599923 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.599936 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.599956 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.599973 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.599982 31304 net.cpp:408] ip <- conv +I0718 09:46:46.599995 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.600257 31304 net.cpp:124] Setting up ip +I0718 09:46:46.600277 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.600288 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.600307 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.600322 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.600332 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.600342 31304 net.cpp:408] loss <- label +I0718 09:46:46.600355 31304 net.cpp:382] loss -> loss +I0718 09:46:46.600375 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.600405 31304 net.cpp:124] Setting up loss +I0718 09:46:46.600415 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.600427 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.600445 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.600454 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.600464 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.600473 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.600483 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.600492 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.600505 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.600549 31304 solver.cpp:57] Solver scaffolding done. +I0718 09:46:46.600737 31304 solver.cpp:45] Initializing solver from parameters: test_iter: 10 test_interval: 10 base_lr: 0.01 @@ -20138,10 +20205,10 @@ momentum: 0.9 weight_decay: 0.0005 snapshot_prefix: "model" -net: "/tmp/tmpk6qfc9et" +net: "/tmp/tmp33qhmbtj" snapshot_after_train: false -I0819 13:44:52.212965 13551 solver.cpp:102] Creating training net from net file: /tmp/tmpk6qfc9et -I0819 13:44:52.213230 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.600826 31304 solver.cpp:102] Creating training net from net file: /tmp/tmp33qhmbtj +I0718 09:46:46.601089 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -20219,46 +20286,46 @@ bottom: "label" top: "loss" } -I0819 13:44:52.213630 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.213673 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.213691 13551 net.cpp:382] data -> data -I0819 13:44:52.213719 13551 net.cpp:382] data -> label -I0819 13:44:52.213752 13551 net.cpp:124] Setting up data -I0819 13:44:52.213764 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.213779 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.213790 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.213799 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.213825 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.213837 13551 net.cpp:408] conv <- data -I0819 13:44:52.213856 13551 net.cpp:382] conv -> conv -I0819 13:44:52.213919 13551 net.cpp:124] Setting up conv -I0819 13:44:52.213934 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.213948 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.213968 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.213986 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.213996 13551 net.cpp:408] ip <- conv -I0819 13:44:52.214011 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.214284 13551 net.cpp:124] Setting up ip -I0819 13:44:52.214306 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.214319 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.214339 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.214354 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.214365 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.214375 13551 net.cpp:408] loss <- label -I0819 13:44:52.214388 13551 net.cpp:382] loss -> loss -I0819 13:44:52.214411 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.214444 13551 net.cpp:124] Setting up loss -I0819 13:44:52.214457 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.214468 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.214488 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.214498 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.214509 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.214519 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.214529 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.214536 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.214552 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.214856 13551 solver.cpp:190] Creating test net (#0) specified by net file: /tmp/tmpk6qfc9et -I0819 13:44:52.214912 13551 net.cpp:53] Initializing net from parameters: +I0718 09:46:46.601311 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.601341 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.601357 31304 net.cpp:382] data -> data +I0718 09:46:46.601382 31304 net.cpp:382] data -> label +I0718 09:46:46.601413 31304 net.cpp:124] Setting up data +I0718 09:46:46.601424 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.601438 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.601449 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.601457 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.601480 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.601490 31304 net.cpp:408] conv <- data +I0718 09:46:46.601509 31304 net.cpp:382] conv -> conv +I0718 09:46:46.601570 31304 net.cpp:124] Setting up conv +I0718 09:46:46.601588 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.601601 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.601622 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.601644 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.601653 31304 net.cpp:408] ip <- conv +I0718 09:46:46.601668 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.601938 31304 net.cpp:124] Setting up ip +I0718 09:46:46.601959 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.601971 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.601991 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.602008 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.602018 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.602030 31304 net.cpp:408] loss <- label +I0718 09:46:46.602044 31304 net.cpp:382] loss -> loss +I0718 09:46:46.602066 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.602098 31304 net.cpp:124] Setting up loss +I0718 09:46:46.602109 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.602120 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.602139 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.602149 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.602156 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.602164 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.602174 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.602181 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.602196 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.602469 31304 solver.cpp:190] Creating test net (#0) specified by net file: /tmp/tmp33qhmbtj +I0718 09:46:46.602522 31304 net.cpp:53] Initializing net from parameters: name: "testnet" force_backward: true state { @@ -20336,77 +20403,77 @@ bottom: "label" top: "loss" } -I0819 13:44:52.215142 13551 layer_factory.hpp:77] Creating layer data -I0819 13:44:52.215169 13551 net.cpp:86] Creating Layer data -I0819 13:44:52.215184 13551 net.cpp:382] data -> data -I0819 13:44:52.215209 13551 net.cpp:382] data -> label -I0819 13:44:52.215237 13551 net.cpp:124] Setting up data -I0819 13:44:52.215248 13551 net.cpp:131] Top shape: 5 2 3 4 (120) -I0819 13:44:52.215263 13551 net.cpp:131] Top shape: 5 1 1 1 (5) -I0819 13:44:52.215276 13551 net.cpp:139] Memory required for data: 500 -I0819 13:44:52.215286 13551 layer_factory.hpp:77] Creating layer conv -I0819 13:44:52.215310 13551 net.cpp:86] Creating Layer conv -I0819 13:44:52.215322 13551 net.cpp:408] conv <- data -I0819 13:44:52.215340 13551 net.cpp:382] conv -> conv -I0819 13:44:52.215404 13551 net.cpp:124] Setting up conv -I0819 13:44:52.215418 13551 net.cpp:131] Top shape: 5 11 8 9 (3960) -I0819 13:44:52.215431 13551 net.cpp:139] Memory required for data: 16340 -I0819 13:44:52.215451 13551 layer_factory.hpp:77] Creating layer ip -I0819 13:44:52.215471 13551 net.cpp:86] Creating Layer ip -I0819 13:44:52.215481 13551 net.cpp:408] ip <- conv -I0819 13:44:52.215497 13551 net.cpp:382] ip -> ip_blob -I0819 13:44:52.215764 13551 net.cpp:124] Setting up ip -I0819 13:44:52.215787 13551 net.cpp:131] Top shape: 5 13 (65) -I0819 13:44:52.215799 13551 net.cpp:139] Memory required for data: 16600 -I0819 13:44:52.215821 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.215837 13551 net.cpp:86] Creating Layer loss -I0819 13:44:52.215847 13551 net.cpp:408] loss <- ip_blob -I0819 13:44:52.215857 13551 net.cpp:408] loss <- label -I0819 13:44:52.215869 13551 net.cpp:382] loss -> loss -I0819 13:44:52.215893 13551 layer_factory.hpp:77] Creating layer loss -I0819 13:44:52.215925 13551 net.cpp:124] Setting up loss -I0819 13:44:52.215937 13551 net.cpp:131] Top shape: (1) -I0819 13:44:52.215948 13551 net.cpp:134] with loss weight 1 -I0819 13:44:52.215970 13551 net.cpp:139] Memory required for data: 16604 -I0819 13:44:52.215979 13551 net.cpp:200] loss needs backward computation. -I0819 13:44:52.215988 13551 net.cpp:200] ip needs backward computation. -I0819 13:44:52.215997 13551 net.cpp:200] conv needs backward computation. -I0819 13:44:52.216007 13551 net.cpp:202] data does not need backward computation. -I0819 13:44:52.216017 13551 net.cpp:244] This network produces output loss -I0819 13:44:52.216032 13551 net.cpp:257] Network initialization done. -I0819 13:44:52.216078 13551 solver.cpp:57] Solver scaffolding done. -I0819 13:44:52.217600 13551 solver.cpp:289] Solving testnet -I0819 13:44:52.217634 13551 solver.cpp:290] Learning Rate Policy: inv -I0819 13:44:52.217702 13551 solver.cpp:347] Iteration 0, Testing net (#0) -I0819 13:44:52.219992 13551 solver.cpp:414] Test net output #0: loss = 69.7827 (* 1 = 69.7827 loss) -I0819 13:44:52.220680 13551 solver.cpp:239] Iteration 0 (0 iter/s, 0.003s/100 iters), loss = 73.6902 -I0819 13:44:52.220724 13551 solver.cpp:258] Train net output #0: loss = 73.6902 (* 1 = 73.6902 loss) -I0819 13:44:52.220743 13551 sgd_solver.cpp:112] Iteration 0, lr = 0.01 -I0819 13:44:52.227147 13551 solver.cpp:347] Iteration 10, Testing net (#0) -I0819 13:44:52.229530 13551 solver.cpp:414] Test net output #0: loss = 74.1233 (* 1 = 74.1233 loss) -I0819 13:44:52.236455 13551 solver.cpp:347] Iteration 20, Testing net (#0) -I0819 13:44:52.238853 13551 solver.cpp:414] Test net output #0: loss = 66.8745 (* 1 = 66.8745 loss) -I0819 13:44:52.245901 13551 solver.cpp:347] Iteration 30, Testing net (#0) -I0819 13:44:52.248174 13551 solver.cpp:414] Test net output #0: loss = 74.3185 (* 1 = 74.3185 loss) -I0819 13:44:52.255239 13551 solver.cpp:347] Iteration 40, Testing net (#0) -I0819 13:44:52.257659 13551 solver.cpp:414] Test net output #0: loss = 62.2477 (* 1 = 62.2477 loss) -I0819 13:44:52.264851 13551 solver.cpp:347] Iteration 50, Testing net (#0) -I0819 13:44:52.267210 13551 solver.cpp:414] Test net output #0: loss = 69.4183 (* 1 = 69.4183 loss) -I0819 13:44:52.274320 13551 solver.cpp:347] Iteration 60, Testing net (#0) -I0819 13:44:52.276616 13551 solver.cpp:414] Test net output #0: loss = 71.7365 (* 1 = 71.7365 loss) -I0819 13:44:52.283659 13551 solver.cpp:347] Iteration 70, Testing net (#0) -I0819 13:44:52.286062 13551 solver.cpp:414] Test net output #0: loss = 72.9062 (* 1 = 72.9062 loss) -I0819 13:44:52.293036 13551 solver.cpp:347] Iteration 80, Testing net (#0) -I0819 13:44:52.295452 13551 solver.cpp:414] Test net output #0: loss = 62.9364 (* 1 = 62.9364 loss) -I0819 13:44:52.302600 13551 solver.cpp:347] Iteration 90, Testing net (#0) -I0819 13:44:52.304929 13551 solver.cpp:414] Test net output #0: loss = 74.7387 (* 1 = 74.7387 loss) -I0819 13:44:52.312142 13551 solver.cpp:327] Iteration 100, loss = 66.1035 -I0819 13:44:52.312206 13551 solver.cpp:347] Iteration 100, Testing net (#0) -I0819 13:44:52.314528 13551 solver.cpp:414] Test net output #0: loss = 70.0752 (* 1 = 70.0752 loss) -I0819 13:44:52.314568 13551 solver.cpp:332] Optimization Done. +I0718 09:46:46.602741 31304 layer_factory.hpp:77] Creating layer data +I0718 09:46:46.602766 31304 net.cpp:86] Creating Layer data +I0718 09:46:46.602780 31304 net.cpp:382] data -> data +I0718 09:46:46.602802 31304 net.cpp:382] data -> label +I0718 09:46:46.602831 31304 net.cpp:124] Setting up data +I0718 09:46:46.602843 31304 net.cpp:131] Top shape: 5 2 3 4 (120) +I0718 09:46:46.602857 31304 net.cpp:131] Top shape: 5 1 1 1 (5) +I0718 09:46:46.602867 31304 net.cpp:139] Memory required for data: 500 +I0718 09:46:46.602876 31304 layer_factory.hpp:77] Creating layer conv +I0718 09:46:46.602898 31304 net.cpp:86] Creating Layer conv +I0718 09:46:46.602908 31304 net.cpp:408] conv <- data +I0718 09:46:46.602924 31304 net.cpp:382] conv -> conv +I0718 09:46:46.602978 31304 net.cpp:124] Setting up conv +I0718 09:46:46.602990 31304 net.cpp:131] Top shape: 5 11 8 9 (3960) +I0718 09:46:46.603001 31304 net.cpp:139] Memory required for data: 16340 +I0718 09:46:46.603021 31304 layer_factory.hpp:77] Creating layer ip +I0718 09:46:46.603039 31304 net.cpp:86] Creating Layer ip +I0718 09:46:46.603050 31304 net.cpp:408] ip <- conv +I0718 09:46:46.603065 31304 net.cpp:382] ip -> ip_blob +I0718 09:46:46.603776 31304 net.cpp:124] Setting up ip +I0718 09:46:46.603808 31304 net.cpp:131] Top shape: 5 13 (65) +I0718 09:46:46.603822 31304 net.cpp:139] Memory required for data: 16600 +I0718 09:46:46.603842 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.603857 31304 net.cpp:86] Creating Layer loss +I0718 09:46:46.603868 31304 net.cpp:408] loss <- ip_blob +I0718 09:46:46.603878 31304 net.cpp:408] loss <- label +I0718 09:46:46.603891 31304 net.cpp:382] loss -> loss +I0718 09:46:46.603912 31304 layer_factory.hpp:77] Creating layer loss +I0718 09:46:46.603942 31304 net.cpp:124] Setting up loss +I0718 09:46:46.603952 31304 net.cpp:131] Top shape: (1) +I0718 09:46:46.603961 31304 net.cpp:134] with loss weight 1 +I0718 09:46:46.603981 31304 net.cpp:139] Memory required for data: 16604 +I0718 09:46:46.603991 31304 net.cpp:200] loss needs backward computation. +I0718 09:46:46.603999 31304 net.cpp:200] ip needs backward computation. +I0718 09:46:46.604007 31304 net.cpp:200] conv needs backward computation. +I0718 09:46:46.604017 31304 net.cpp:202] data does not need backward computation. +I0718 09:46:46.604025 31304 net.cpp:244] This network produces output loss +I0718 09:46:46.604038 31304 net.cpp:257] Network initialization done. +I0718 09:46:46.604087 31304 solver.cpp:57] Solver scaffolding done. +I0718 09:46:46.605484 31304 solver.cpp:289] Solving testnet +I0718 09:46:46.605512 31304 solver.cpp:290] Learning Rate Policy: inv +I0718 09:46:46.605561 31304 solver.cpp:347] Iteration 0, Testing net (#0) +I0718 09:46:46.608109 31304 solver.cpp:414] Test net output #0: loss = 65.0524 (* 1 = 65.0524 loss) +I0718 09:46:46.608795 31304 solver.cpp:239] Iteration 0 (7.69473e-24 iter/s, 0.003s/100 iters), loss = 69.8692 +I0718 09:46:46.608841 31304 solver.cpp:258] Train net output #0: loss = 69.8692 (* 1 = 69.8692 loss) +I0718 09:46:46.608860 31304 sgd_solver.cpp:112] Iteration 0, lr = 0.01 +I0718 09:46:46.615406 31304 solver.cpp:347] Iteration 10, Testing net (#0) +I0718 09:46:46.617673 31304 solver.cpp:414] Test net output #0: loss = 86.0826 (* 1 = 86.0826 loss) +I0718 09:46:46.625038 31304 solver.cpp:347] Iteration 20, Testing net (#0) +I0718 09:46:46.627669 31304 solver.cpp:414] Test net output #0: loss = 86.2211 (* 1 = 86.2211 loss) +I0718 09:46:46.634914 31304 solver.cpp:347] Iteration 30, Testing net (#0) +I0718 09:46:46.637472 31304 solver.cpp:414] Test net output #0: loss = 80.0753 (* 1 = 80.0753 loss) +I0718 09:46:46.644766 31304 solver.cpp:347] Iteration 40, Testing net (#0) +I0718 09:46:46.647025 31304 solver.cpp:414] Test net output #0: loss = 82.6375 (* 1 = 82.6375 loss) +I0718 09:46:46.654328 31304 solver.cpp:347] Iteration 50, Testing net (#0) +I0718 09:46:46.656716 31304 solver.cpp:414] Test net output #0: loss = 84.2571 (* 1 = 84.2571 loss) +I0718 09:46:46.664024 31304 solver.cpp:347] Iteration 60, Testing net (#0) +I0718 09:46:46.666293 31304 solver.cpp:414] Test net output #0: loss = 81.9866 (* 1 = 81.9866 loss) +I0718 09:46:46.673652 31304 solver.cpp:347] Iteration 70, Testing net (#0) +I0718 09:46:46.676280 31304 solver.cpp:414] Test net output #0: loss = 78.8137 (* 1 = 78.8137 loss) +I0718 09:46:46.683569 31304 solver.cpp:347] Iteration 80, Testing net (#0) +I0718 09:46:46.685855 31304 solver.cpp:414] Test net output #0: loss = 75.7248 (* 1 = 75.7248 loss) +I0718 09:46:46.693228 31304 solver.cpp:347] Iteration 90, Testing net (#0) +I0718 09:46:46.695641 31304 solver.cpp:414] Test net output #0: loss = 82.1313 (* 1 = 82.1313 loss) +I0718 09:46:46.702950 31304 solver.cpp:327] Iteration 100, loss = 63.4808 +I0718 09:46:46.703045 31304 solver.cpp:347] Iteration 100, Testing net (#0) +I0718 09:46:46.705490 31304 solver.cpp:414] Test net output #0: loss = 81.616 (* 1 = 81.616 loss) +I0718 09:46:46.705544 31304 solver.cpp:332] Optimization Done. . ---------------------------------------------------------------------- -Ran 52 tests in 29.802s +Ran 52 tests in 30.256s OK make[5]: Leaving directory '/build/caffe-1.0.0+git20180821.99bd997/caffe_cpu_build' @@ -20636,11 +20703,11 @@ dh_shlibdeps -a -O-Scmake dh_installdeb -O-Scmake dh_gencontrol -O-Scmake -dpkg-gencontrol: warning: package libcaffe-dev: substitution variable ${python3:Depends} unused, but is defined -dpkg-gencontrol: warning: package libcaffe1: substitution variable ${python3:Depends} unused, but is defined dpkg-gencontrol: warning: package python3-caffe: substitution variable ${python3:Provides} unused, but is defined dpkg-gencontrol: warning: package python3-caffe: substitution variable ${python3:Versions} unused, but is defined +dpkg-gencontrol: warning: package libcaffe1: substitution variable ${python3:Depends} unused, but is defined dpkg-gencontrol: warning: package caffe: substitution variable ${python3:Depends} unused, but is defined +dpkg-gencontrol: warning: package libcaffe-dev: substitution variable ${python3:Depends} unused, but is defined dpkg-gencontrol: warning: package libcaffe1: substitution variable ${python3:Depends} unused, but is defined dpkg-gencontrol: warning: package python3-caffe: substitution variable ${python3:Provides} unused, but is defined dpkg-gencontrol: warning: package python3-caffe: substitution variable ${python3:Versions} unused, but is defined @@ -20649,13 +20716,13 @@ dh_md5sums -O-Scmake dh_builddeb -O-Scmake dpkg-deb: building package 'caffe-dbgsym' in '../caffe-dbgsym_1.0.0+git20180821.99bd997-8_arm64.deb'. -dpkg-deb: building package 'libcaffe1-dbgsym' in '../libcaffe1-dbgsym_1.0.0+git20180821.99bd997-8_arm64.deb'. -dpkg-deb: building package 'libcaffe1' in '../libcaffe1_1.0.0+git20180821.99bd997-8_arm64.deb'. dpkg-deb: building package 'caffe' in '../caffe_1.0.0+git20180821.99bd997-8_arm64.deb'. -dpkg-deb: building package 'python3-caffe' in '../python3-caffe_1.0.0+git20180821.99bd997-8_arm64.deb'. -dpkg-deb: building package 'caffe-doc' in '../caffe-doc_1.0.0+git20180821.99bd997-8_all.deb'. +dpkg-deb: building package 'libcaffe1' in '../libcaffe1_1.0.0+git20180821.99bd997-8_arm64.deb'. dpkg-deb: building package 'libcaffe-dev' in '../libcaffe-dev_1.0.0+git20180821.99bd997-8_arm64.deb'. dpkg-deb: building package 'python3-caffe-dbgsym' in '../python3-caffe-dbgsym_1.0.0+git20180821.99bd997-8_arm64.deb'. +dpkg-deb: building package 'caffe-doc' in '../caffe-doc_1.0.0+git20180821.99bd997-8_all.deb'. +dpkg-deb: building package 'libcaffe1-dbgsym' in '../libcaffe1-dbgsym_1.0.0+git20180821.99bd997-8_arm64.deb'. +dpkg-deb: building package 'python3-caffe' in '../python3-caffe_1.0.0+git20180821.99bd997-8_arm64.deb'. dpkg-genbuildinfo --build=binary dpkg-genchanges --build=binary >../caffe_1.0.0+git20180821.99bd997-8_arm64.changes dpkg-genchanges: info: binary-only upload (no source code included) @@ -20663,12 +20730,14 @@ dpkg-buildpackage: info: binary-only upload (no source included) dpkg-genchanges: info: not including original source code in upload I: copying local configuration +I: user script /srv/workspace/pbuilder/3330/tmp/hooks/B01_cleanup starting +I: user script /srv/workspace/pbuilder/3330/tmp/hooks/B01_cleanup finished I: unmounting dev/ptmx filesystem I: unmounting dev/pts filesystem I: unmounting dev/shm filesystem I: unmounting proc filesystem I: unmounting sys filesystem I: cleaning the build env -I: removing directory /srv/workspace/pbuilder/29121 and its subdirectories -I: Current time: Fri Aug 19 13:47:09 -12 2022 -I: pbuilder-time-stamp: 1660960029 +I: removing directory /srv/workspace/pbuilder/3330 and its subdirectories +I: Current time: Sun Jul 18 09:49:19 +14 2021 +I: pbuilder-time-stamp: 1626551359