diff --git a/README.md b/README.md index 3588f14..1b3c81d 100644 --- a/README.md +++ b/README.md @@ -23,12 +23,7 @@ and to the video presentation [here](https://www.youtube.com/watch?v=Q7Q9o7ywXx8 ## Install TimeSHAP -##### Via Pip -``` -pip install timeshap -``` - -##### Via Github +#### [Recommended] Via Github Clone the repository into a local directory using: ``` git clone https://github.com/feedzai/timeshap.git @@ -41,8 +36,12 @@ cd timeshap pip install . ``` +#### Via Pip +``` +pip install timeshap +``` -##### Test your installation +#### Test your installation Start a Python session in your terminal using ``` diff --git a/notebooks/AReM/AReM.ipynb b/notebooks/AReM/AReM.ipynb index 273dbab..db2d3bf 100644 --- a/notebooks/AReM/AReM.ipynb +++ b/notebooks/AReM/AReM.ipynb @@ -69,7 +69,7 @@ { "data": { "text/plain": [ - "'1.0.2'" + "'1.1.0'" ] }, "execution_count": 2, @@ -124,80 +124,80 @@ "name": "stdout", "output_type": "stream", "text": [ - "sitting/dataset5.csv ------ (480, 7)\n", - "sitting/dataset11.csv ------ (480, 7)\n", - "sitting/dataset3.csv ------ (480, 7)\n", - "sitting/dataset10.csv ------ (480, 7)\n", - "sitting/dataset12.csv ------ (480, 7)\n", - "sitting/dataset4.csv ------ (480, 7)\n", - "sitting/dataset13.csv ------ (480, 7)\n", - "sitting/dataset9.csv ------ (480, 7)\n", - "sitting/dataset2.csv ------ (480, 7)\n", - "sitting/dataset15.csv ------ (480, 7)\n", - "sitting/dataset1.csv ------ (480, 7)\n", - "sitting/dataset14.csv ------ (480, 7)\n", - "sitting/dataset6.csv ------ (480, 7)\n", - "sitting/dataset7.csv ------ (480, 7)\n", - "standing/dataset5.csv ------ (480, 7)\n", - "standing/dataset11.csv ------ (480, 7)\n", - "standing/dataset3.csv ------ (480, 7)\n", - "standing/dataset8.csv ------ (480, 7)\n", - "standing/dataset10.csv ------ (480, 7)\n", - "standing/dataset12.csv ------ (480, 7)\n", - "standing/dataset4.csv ------ (480, 7)\n", - "standing/dataset13.csv ------ (480, 7)\n", - "standing/dataset9.csv ------ (480, 7)\n", - "standing/dataset2.csv ------ (480, 7)\n", - "standing/dataset15.csv ------ (480, 7)\n", - "standing/dataset1.csv ------ (480, 7)\n", - "standing/dataset14.csv ------ (480, 7)\n", - "standing/dataset6.csv ------ (480, 7)\n", - "standing/dataset7.csv ------ (480, 7)\n", + "walking/dataset7.csv ------ (480, 7)\n", + "walking/dataset6.csv ------ (480, 7)\n", + "walking/dataset4.csv ------ (480, 7)\n", "walking/dataset5.csv ------ (480, 7)\n", - "walking/dataset11.csv ------ (480, 7)\n", + "walking/dataset1.csv ------ (480, 7)\n", + "walking/dataset2.csv ------ (480, 7)\n", "walking/dataset3.csv ------ (480, 7)\n", - "walking/dataset8.csv ------ (480, 7)\n", "walking/dataset10.csv ------ (480, 7)\n", - "walking/dataset12.csv ------ (480, 7)\n", - "walking/dataset4.csv ------ (480, 7)\n", + "walking/dataset11.csv ------ (480, 7)\n", "walking/dataset13.csv ------ (480, 7)\n", - "walking/dataset9.csv ------ (480, 7)\n", - "walking/dataset2.csv ------ (480, 7)\n", + "walking/dataset12.csv ------ (480, 7)\n", "walking/dataset15.csv ------ (480, 7)\n", - "walking/dataset1.csv ------ (480, 7)\n", "walking/dataset14.csv ------ (480, 7)\n", - "walking/dataset6.csv ------ (480, 7)\n", - "walking/dataset7.csv ------ (480, 7)\n", - "cycling/dataset5.csv ------ (480, 7)\n", - "cycling/dataset11.csv ------ (480, 7)\n", - "cycling/dataset3.csv ------ (480, 7)\n", - "cycling/dataset8.csv ------ (480, 7)\n", - "cycling/dataset10.csv ------ (480, 7)\n", - "cycling/dataset12.csv ------ (480, 7)\n", - "cycling/dataset4.csv ------ (480, 7)\n", - "cycling/dataset13.csv ------ (480, 7)\n", - "cycling/dataset9.csv ------ (480, 7)\n", - "cycling/dataset2.csv ------ (480, 7)\n", - "cycling/dataset15.csv ------ (480, 7)\n", - "cycling/dataset1.csv ------ (480, 7)\n", - "cycling/dataset14.csv ------ (480, 7)\n", - "cycling/dataset6.csv ------ (480, 7)\n", - "cycling/dataset7.csv ------ (480, 7)\n", + "walking/dataset8.csv ------ (480, 7)\n", + "walking/dataset9.csv ------ (480, 7)\n", + "standing/dataset7.csv ------ (480, 7)\n", + "standing/dataset6.csv ------ (480, 7)\n", + "standing/dataset4.csv ------ (480, 7)\n", + "standing/dataset5.csv ------ (480, 7)\n", + "standing/dataset1.csv ------ (480, 7)\n", + "standing/dataset2.csv ------ (480, 7)\n", + "standing/dataset3.csv ------ (480, 7)\n", + "standing/dataset10.csv ------ (480, 7)\n", + "standing/dataset11.csv ------ (480, 7)\n", + "standing/dataset13.csv ------ (480, 7)\n", + "standing/dataset12.csv ------ (480, 7)\n", + "standing/dataset15.csv ------ (480, 7)\n", + "standing/dataset14.csv ------ (480, 7)\n", + "standing/dataset8.csv ------ (480, 7)\n", + "standing/dataset9.csv ------ (480, 7)\n", + "sitting/dataset7.csv ------ (480, 7)\n", + "sitting/dataset6.csv ------ (480, 7)\n", + "sitting/dataset4.csv ------ (480, 7)\n", + "sitting/dataset5.csv ------ (480, 7)\n", + "sitting/dataset1.csv ------ (480, 7)\n", + "sitting/dataset2.csv ------ (480, 7)\n", + "sitting/dataset3.csv ------ (480, 7)\n", + "sitting/dataset10.csv ------ (480, 7)\n", + "sitting/dataset11.csv ------ (480, 7)\n", + "sitting/dataset13.csv ------ (480, 7)\n", + "sitting/dataset12.csv ------ (480, 7)\n", + "sitting/dataset15.csv ------ (480, 7)\n", + "sitting/dataset14.csv ------ (480, 7)\n", + "sitting/dataset9.csv ------ (480, 7)\n", + "lying/dataset7.csv ------ (480, 7)\n", + "lying/dataset6.csv ------ (480, 7)\n", + "lying/dataset4.csv ------ (480, 7)\n", "lying/dataset5.csv ------ (480, 7)\n", - "lying/dataset11.csv ------ (480, 7)\n", + "lying/dataset1.csv ------ (480, 7)\n", + "lying/dataset2.csv ------ (480, 7)\n", "lying/dataset3.csv ------ (480, 7)\n", - "lying/dataset8.csv ------ (480, 7)\n", "lying/dataset10.csv ------ (480, 7)\n", - "lying/dataset12.csv ------ (480, 7)\n", - "lying/dataset4.csv ------ (480, 7)\n", + "lying/dataset11.csv ------ (480, 7)\n", "lying/dataset13.csv ------ (480, 7)\n", - "lying/dataset9.csv ------ (480, 7)\n", - "lying/dataset2.csv ------ (480, 7)\n", + "lying/dataset12.csv ------ (480, 7)\n", "lying/dataset15.csv ------ (480, 7)\n", - "lying/dataset1.csv ------ (480, 7)\n", "lying/dataset14.csv ------ (480, 7)\n", - "lying/dataset6.csv ------ (480, 7)\n", - "lying/dataset7.csv ------ (480, 7)\n" + "lying/dataset8.csv ------ (480, 7)\n", + "lying/dataset9.csv ------ (480, 7)\n", + "cycling/dataset7.csv ------ (480, 7)\n", + "cycling/dataset6.csv ------ (480, 7)\n", + "cycling/dataset4.csv ------ (480, 7)\n", + "cycling/dataset5.csv ------ (480, 7)\n", + "cycling/dataset1.csv ------ (480, 7)\n", + "cycling/dataset2.csv ------ (480, 7)\n", + "cycling/dataset3.csv ------ (480, 7)\n", + "cycling/dataset10.csv ------ (480, 7)\n", + "cycling/dataset11.csv ------ (480, 7)\n", + "cycling/dataset13.csv ------ (480, 7)\n", + "cycling/dataset12.csv ------ (480, 7)\n", + "cycling/dataset15.csv ------ (480, 7)\n", + "cycling/dataset14.csv ------ (480, 7)\n", + "cycling/dataset8.csv ------ (480, 7)\n", + "cycling/dataset9.csv ------ (480, 7)\n" ] } ], @@ -524,125 +524,67 @@ "name": "stderr", "output_type": "stream", "text": [ - " 12%|█▎ | 1/8 [00:02<00:15, 2.23s/it]" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Train loss: 0.6997937560081482 --- Test loss 0.5927574634552002 \n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "\r", - " 25%|██▌ | 2/8 [00:04<00:12, 2.01s/it]" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Train loss: 0.5918577313423157 --- Test loss 0.4946227967739105 \n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "\r", - " 38%|███▊ | 3/8 [00:05<00:08, 1.72s/it]" + " 25%|██▌ | 2/8 [00:00<00:01, 5.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Train loss: 0.49630260467529297 --- Test loss 0.40795645117759705 \n" + "Train loss: 0.7000859975814819 --- Test loss 0.5926060080528259 \n", + "Train loss: 0.593891441822052 --- Test loss 0.4921618103981018 \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ - "\r", - " 50%|█████ | 4/8 [00:07<00:06, 1.69s/it]" + " 50%|█████ | 4/8 [00:00<00:00, 5.78it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Train loss: 0.4160643219947815 --- Test loss 0.343948632478714 \n" + "Train loss: 0.4992573857307434 --- Test loss 0.40382787585258484 \n", + "Train loss: 0.41805166006088257 --- Test loss 0.34258148074150085 \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ - "\r", - " 62%|██████▎ | 5/8 [00:10<00:06, 2.16s/it]" + " 75%|███████▌ | 6/8 [00:01<00:00, 5.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Train loss: 0.3660268187522888 --- Test loss 0.2917708158493042 \n" + "Train loss: 0.3631644546985626 --- Test loss 0.30080685019493103 \n", + "Train loss: 0.32556360960006714 --- Test loss 0.260188490152359 \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ - "\r", - " 75%|███████▌ | 6/8 [00:12<00:04, 2.37s/it]" + "100%|██████████| 8/8 [00:01<00:00, 5.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Train loss: 0.334475576877594 --- Test loss 0.2376544028520584 \n" + "Train loss: 0.28500068187713623 --- Test loss 0.22024422883987427 \n", + "Train loss: 0.23866622149944305 --- Test loss 0.18811321258544922 \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ - "\r", - " 88%|████████▊ | 7/8 [00:16<00:02, 2.69s/it]" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Train loss: 0.29900506138801575 --- Test loss 0.1931639164686203 \n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "100%|██████████| 8/8 [00:19<00:00, 2.43s/it]" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Train loss: 0.25920480489730835 --- Test loss 0.17529389262199402 \n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "\n" + "100%|██████████| 8/8 [00:01<00:00, 5.64it/s]\n" ] } ], @@ -764,12 +706,12 @@ "
\n", "