Name | sema-toolchain JSON |
Version |
0.0.11
JSON |
| download |
home_page | None |
Summary | Python symbolic execution package |
upload_time | 2024-05-20 09:11:11 |
maintainer | None |
docs_url | None |
author | None |
requires_python | >=3.8 |
license | BSD 2-Clause License Copyright (c) 2022, UCL-Cybersecurity team All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: 1. Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. 2. Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
keywords |
scdg
binary
symbolic
analysis
|
VCS |
|
bugtrack_url |
|
requirements |
No requirements were recorded.
|
Travis-CI |
No Travis.
|
coveralls test coverage |
No coveralls.
|
# SEMA - ToolChain using Symbolic Execution for Malware Analysis.
```
██████ ▓█████ ███▄ ▄███▓ ▄▄▄
▒██ ▒ ▓█ ▀ ▓██▒▀█▀ ██▒▒████▄
░ ▓██▄ ▒███ ▓██ ▓██░▒██ ▀█▄
▒ ██▒▒▓█ ▄ ▒██ ▒██ ░██▄▄▄▄██
▒██████▒▒░▒████▒▒██▒ ░██▒ ▓█ ▓██▒
▒ ▒▓▒ ▒ ░░░ ▒░ ░░ ▒░ ░ ░ ▒▒ ▓▒█░
░ ░▒ ░ ░ ░ ░ ░░ ░ ░ ▒ ▒▒ ░
░ ░ ░ ░ ░ ░ ░ ▒
░ ░ ░ ░ ░ ░
```
# Table of Contents
1. [Architecture](#architecture)
- [Toolchain Architecture](#toolchain-architecture)
2. [Recommended Installation and Usage](#page_with_curl-recommended-installation-and-usage)
3. [Dockerhub Installation and Usage](#page_with_curl-dockerhub-installation)
4. [Pypi Installation and Usage](#page_with_curl-pypi-installation-and-usage)
5. [Credentials](#page_with_curl-credentials)
Architecture
====
<a name="architecture"></a>
### Toolchain architecture
<a name="toolchain-architecture"></a>
Our toolchain is represented in the following figure and works as follows:
- A collection of labelled binaries from different malware families is collected and used as the input of the toolchain.
- **Angr**, a framework for symbolic execution, is used to execute binaries symbolically and extract execution traces. For this purpose, different heuristics have been developed to optimize symbolic execution.
- Several execution traces (i.e., API calls used and their arguments) corresponding to one binary are extracted with Angr and gathered together using several graph heuristics to construct a SCDG.
- These resulting SCDGs are then used as input to graph mining to extract common graphs between SCDGs of the same family and create a signature.
- Finally, when a new sample has to be classified, its SCDG is built and compared with SCDGs of known families using a simple similarity metric.
This repository contains a first version of a SCDG extractor. During the symbolic analysis of a binary, all system calls and their arguments found are recorded. After some stop conditions for symbolic analysis, a graph is built as follows: Nodes are system calls recorded, edges show that some arguments are shared between calls.
When a new sample has to be evaluated, its SCDG is first built as described previously. Then, `gspan` is applied to extract the biggest common subgraph and a similarity score is evaluated to decide if the graph is considered as part of the family or not. The similarity score `S` between graph `G'` and `G''` is computed as follows:
Since `G''` is a subgraph of `G'`, this is calculating how much `G'` appears in `G''`.
Another classifier we use is the Support Vector Machine (`SVM`) with INRIA graph kernel or the Weisfeiler-Lehman extension graph kernel.
A web application is available and is called SemaWebApp. It allows to manage the launch of experiments on SemaSCDG and/or SemaClassifier.
#### Main depencies:
* Python 3.8
* Docker >=26.1.3 , docker buildx, Docker Compose >=v2.27.0
* radare2
* libvirt-dev, libgraphviz-dev, wheel
#### Interesting links
- [Angr](https://angr.io/)
- [Bazaar Abuse](https://bazaar.abuse.ch/)
- [Docker Installation on Ubuntu](https://docs.docker.com/engine/install/ubuntu/)
#### Extracting database
To extract the database, use the following commands:
```bash
cd databases/Binaries
./extract_deploy_db.sh
```
Password for archive is "infected". Warning : it contains real samples of malwares.
#### Compressing database
To compress the database, use the following commands:
```bash
#To zip back the test database
cd databases/Binaries
./compress_db.sh
```
**Pypi installation and usage**
====
<a name="pypi-installation-and-usage"></a>
To use the toolchain without docker container by using the Pypi package to install dependencies, use :
```bash
pip install sema-toolchain
```
After cloning the git you can then use the toolchain without docker
## Pypy3 usage
Pypy3 can be used to launch experiments, make sure to install pypy3 :
```bash
sudo add-apt-repository ppa:pypy/ppa
sudo apt update
sudo apt install pypy3
```
Then install the dependecies on pypy3 :
```bash
pypy3 -m pip install -r /sema_scdg/requirements_pypy.txt
```
## How to use ?
### Use SemaSCDG
To run experiments, run :
```bash
python3 sema_scdg/application/SemaSCDG.py sema_scdg/application/configs/config.ini
```
Or if you want to use pypy3:
```bash
pypy3 sema_scdg/application/SemaSCDG.py sema_scdg/application/configs/config.ini
```
#### Configuration files
The parameters are put in a configuration file : `configs/config.ini`. Feel free to modify it or create new configuration files to run different experiments.
The output of the SCDG are put into `database/SCDG/runs/` by default. If you are not using volumes and want to save some runs from the container to your host machine, use :
```bash
make save-scdg-runs ARGS=PATH
```
#### Parameters description
SCDG module arguments
```
expl_method:
DFS Depth First Search
BFS Breadth First Search
CDFS Coverage Depth-First Search Strategy (Default)
CBFS Coverage Breadth First Search
graph_output:
gs .GS format
json .JSON format
EMPTY if left empty then build on all available format
packing_type:
symbion Concolic unpacking method (linux | windows [in progress])
unipacker Emulation unpacking method (windows only)
SCDG exploration techniques parameters:
jump_it Number of iteration allowed for a symbolic loop (default : 3)
max_in_pause_stach Number of states allowed in pause stash (default : 200)
max_step Maximum number of steps allowed for a state (default : 50 000)
max_end_state Number of deadended state required to stop (default : 600)
max_simul_state Number of simultaneous states we explore with simulation manager (default : 5)
Binary parameters:
n_args Number of symbolic arguments given to the binary (default : 0)
loop_counter_concrete How many times a loop can loop (default : 10240)
count_block_enable Enable the count of visited blocks and instructions
sim_file Create SimFile
entry_addr Entry address of the binary
SCDG creation parameter:
min_size Minimum size required for a trace to be used in SCDG (default : 3)
disjoint_union Do we merge traces or use disjoint union ? (default : merge)
not_comp_args Do we compare arguments to add new nodes when building graph ? (default : comparison enabled)
three_edges Do we use the three-edges strategy ? (default : False)
not_ignore_zero Do we ignore zero when building graph ? (default : Discard zero)
keep_inter_SCDG Keep intermediate SCDG in file (default : False)
eval_time TODO
Global parameter:
concrete_target_is_local Use a local GDB server instead of using cuckoo (default : False)
print_syscall Print the syscall found
csv_file Name of the csv to save the experiment data
plugin_enable Enable the plugins set to true in the config.ini file
approximate Symbolic approximation
is_packed Is the binary packed ? (default : False, not yet supported)
timeout Timeout in seconds before ending extraction (default : 600)
string_resolve Do we try to resolv references of string (default : True)
log_level Level of log, can be INFO, DEBUG, WARNING, ERROR (default : INFO)
family Family of the malware (default : Unknown)
exp_dir Name of the directory to save SCDG extracted (default : Default)
binary_path Relative path to the binary or directory (has to be in the database folder)
fast_main Jump directly into the main function
Plugins:
plugin_env_var Enable the env_var plugin
plugin_locale_info Enable the locale_info plugin
plugin_resources Enable the resources plugin
plugin_widechar Enable the widechar plugin
plugin_registery Enable the registery plugin
plugin_atom Enable the atom plugin
plugin_thread Enable the thread plugin
plugin_track_command Enable the track_command plugin
plugin_ioc_report Enable the ioc_report plugin
plugin_hooks Enable the hooks plugin
```
To know the details of the angr options see [Angr documentation](https://docs.angr.io/en/latest/appendix/options.html)
You also have a script `MergeGspan.py` in `sema_scdg/application/helper` which could merge all `.gs` from a directory into only one file.
#### Run multiple experiments automatically
If you wish to run multiple experiments with different configuration files, the script `multiple_experiments.sh` is available. When being in the folder containing SemaSCDG.py :
```bash
# To show usage
./multiple_experiments.sh -h
# Run example
./multiple_experiments.sh -m python3 -c configs/config1.ini configs/config2.ini
```
#### Tests
To run the test :
```bash
python3 scdg_tests.py test_data/config_test.ini
```
#### Tutorial
There is a jupyter notebook providing a tutorial on how to use the scdg. To launch it, run
```bash
jupyter notebook --ip=0.0.0.0 --port=5001 --no-browser --allow-root --NotebookApp.token=''
```
and visit `http://127.0.0.1:5001/tree` on your browser. Go to `/Tutorial` and open the jupyter notebook.
### Use SemaClassifier
Just run the script :
```
python3 SemaClassifier.py FOLDER/FILE
usage: update_readme_usage.py [-h] [--threshold THRESHOLD] [--biggest_subgraph BIGGEST_SUBGRAPH] [--support SUPPORT] [--ctimeout CTIMEOUT] [--epoch EPOCH] [--sepoch SEPOCH]
[--data_scale DATA_SCALE] [--vector_size VECTOR_SIZE] [--batch_size BATCH_SIZE] (--classification | --detection) (--wl | --inria | --dl | --gspan)
[--bancteian] [--delf] [--FeakerStealer] [--gandcrab] [--ircbot] [--lamer] [--nitol] [--RedLineStealer] [--sfone] [--sillyp2p] [--simbot]
[--Sodinokibi] [--sytro] [--upatre] [--wabot] [--RemcosRAT] [--verbose_classifier] [--train] [--nthread NTHREAD]
binaries
Classification module arguments
optional arguments:
-h, --help show this help message and exit
--classification By malware family
--detection Cleanware vs Malware
--wl TODO
--inria TODO
--dl TODO
--gspan TODOe
Global classifiers parameters:
--threshold THRESHOLD
Threshold used for the classifier [0..1] (default : 0.45)
Gspan options:
--biggest_subgraph BIGGEST_SUBGRAPH
Biggest subgraph consider for Gspan (default: 5)
--support SUPPORT Support used for the gpsan classifier [0..1] (default : 0.75)
--ctimeout CTIMEOUT Timeout for gspan classifier (default : 3sec)
Deep Learning options:
--epoch EPOCH Only for deep learning model: number of epoch (default: 5) Always 1 for FL model
--sepoch SEPOCH Only for deep learning model: starting epoch (default: 1)
--data_scale DATA_SCALE
Only for deep learning model: data scale value (default: 0.9)
--vector_size VECTOR_SIZE
Only for deep learning model: Size of the vector used (default: 4)
--batch_size BATCH_SIZE
Only for deep learning model: Batch size for the model (default: 1)
Malware familly:
--bancteian
--delf
--FeakerStealer
--gandcrab
--ircbot
--lamer
--nitol
--RedLineStealer
--sfone
--sillyp2p
--simbot
--Sodinokibi
--sytro
--upatre
--wabot
--RemcosRAT
Global parameter:
--verbose_classifier Verbose output during train/classification (default : False)
--train Launch training process, else classify/detect new sample with previously computed model
--nthread NTHREAD Number of thread used (default: max)
binaries Name of the folder containing binary'signatures to analyze (Default: output/save-SCDG/, only that for ToolChain)
```
#### Example
This will train models for input dataset
```bash
python3 SemaClassifier.py --train output/save-SCDG/
```
This will classify input dataset based on previously computed models
```bash
python3 SemaClassifier.py output/test-set/
```
#### Tests
To run the classifier tests :
```bash
python3 classifier_tests.py configs/config_test.ini
```
Credentials
====
<a name="credentials"></a>
Main authors of the projects:
* **Charles-Henry Bertrand Van Ouytsel** (UCLouvain)
* **Christophe Crochet** (UCLouvain)
* **Khanh Huu The Dam** (UCLouvain)
* **Oreins Manon** (UCLouvain)
Under the supervision and with the support of **Fabrizio Biondi** (Avast)
Under the supervision and with the support of our professor **Axel Legay** (UCLouvain) (:heart:)
Raw data
{
"_id": null,
"home_page": null,
"name": "sema-toolchain",
"maintainer": null,
"docs_url": null,
"requires_python": ">=3.8",
"maintainer_email": null,
"keywords": "scdg, binary, symbolic, analysis",
"author": null,
"author_email": "Manon-Oreins <manon.oreins@gmail.com>",
"download_url": "https://files.pythonhosted.org/packages/85/30/031de95490470d18e522643000b04498edf89c9ffec556f534c3be599e9b/sema_toolchain-0.0.11.tar.gz",
"platform": null,
"description": "# SEMA - ToolChain using Symbolic Execution for Malware Analysis.\n\n```\n \u2588\u2588\u2588\u2588\u2588\u2588 \u2593\u2588\u2588\u2588\u2588\u2588 \u2588\u2588\u2588\u2584 \u2584\u2588\u2588\u2588\u2593 \u2584\u2584\u2584\n\u2592\u2588\u2588 \u2592 \u2593\u2588 \u2580 \u2593\u2588\u2588\u2592\u2580\u2588\u2580 \u2588\u2588\u2592\u2592\u2588\u2588\u2588\u2588\u2584\n\u2591 \u2593\u2588\u2588\u2584 \u2592\u2588\u2588\u2588 \u2593\u2588\u2588 \u2593\u2588\u2588\u2591\u2592\u2588\u2588 \u2580\u2588\u2584\n \u2592 \u2588\u2588\u2592\u2592\u2593\u2588 \u2584 \u2592\u2588\u2588 \u2592\u2588\u2588 \u2591\u2588\u2588\u2584\u2584\u2584\u2584\u2588\u2588\n\u2592\u2588\u2588\u2588\u2588\u2588\u2588\u2592\u2592\u2591\u2592\u2588\u2588\u2588\u2588\u2592\u2592\u2588\u2588\u2592 \u2591\u2588\u2588\u2592 \u2593\u2588 \u2593\u2588\u2588\u2592\n\u2592 \u2592\u2593\u2592 \u2592 \u2591\u2591\u2591 \u2592\u2591 \u2591\u2591 \u2592\u2591 \u2591 \u2591 \u2592\u2592 \u2593\u2592\u2588\u2591\n\u2591 \u2591\u2592 \u2591 \u2591 \u2591 \u2591 \u2591\u2591 \u2591 \u2591 \u2592 \u2592\u2592 \u2591\n\u2591 \u2591 \u2591 \u2591 \u2591 \u2591 \u2591 \u2592\n \u2591 \u2591 \u2591 \u2591 \u2591 \u2591\n\n```\n\n\n# Table of Contents\n1. [Architecture](#architecture)\n - [Toolchain Architecture](#toolchain-architecture)\n2. [Recommended Installation and Usage](#page_with_curl-recommended-installation-and-usage)\n3. [Dockerhub Installation and Usage](#page_with_curl-dockerhub-installation)\n4. [Pypi Installation and Usage](#page_with_curl-pypi-installation-and-usage)\n5. [Credentials](#page_with_curl-credentials)\n\nArchitecture\n====\n<a name=\"architecture\"></a>\n\n### Toolchain architecture\n<a name=\"toolchain-architecture\"></a>\n\nOur toolchain is represented in the following figure and works as follows:\n\n- A collection of labelled binaries from different malware families is collected and used as the input of the toolchain.\n- **Angr**, a framework for symbolic execution, is used to execute binaries symbolically and extract execution traces. For this purpose, different heuristics have been developed to optimize symbolic execution.\n- Several execution traces (i.e., API calls used and their arguments) corresponding to one binary are extracted with Angr and gathered together using several graph heuristics to construct a SCDG.\n- These resulting SCDGs are then used as input to graph mining to extract common graphs between SCDGs of the same family and create a signature.\n- Finally, when a new sample has to be classified, its SCDG is built and compared with SCDGs of known families using a simple similarity metric.\n\nThis repository contains a first version of a SCDG extractor. During the symbolic analysis of a binary, all system calls and their arguments found are recorded. After some stop conditions for symbolic analysis, a graph is built as follows: Nodes are system calls recorded, edges show that some arguments are shared between calls.\n\nWhen a new sample has to be evaluated, its SCDG is first built as described previously. Then, `gspan` is applied to extract the biggest common subgraph and a similarity score is evaluated to decide if the graph is considered as part of the family or not. The similarity score `S` between graph `G'` and `G''` is computed as follows:\nSince `G''` is a subgraph of `G'`, this is calculating how much `G'` appears in `G''`.\nAnother classifier we use is the Support Vector Machine (`SVM`) with INRIA graph kernel or the Weisfeiler-Lehman extension graph kernel.\n\nA web application is available and is called SemaWebApp. It allows to manage the launch of experiments on SemaSCDG and/or SemaClassifier.\n\n#### Main depencies:\n\n * Python 3.8\n\n * Docker >=26.1.3 , docker buildx, Docker Compose >=v2.27.0\n\n * radare2\n\n * libvirt-dev, libgraphviz-dev, wheel\n\n#### Interesting links\n\n\n- [Angr](https://angr.io/)\n- [Bazaar Abuse](https://bazaar.abuse.ch/)\n- [Docker Installation on Ubuntu](https://docs.docker.com/engine/install/ubuntu/)\n\n\n#### Extracting database\n\nTo extract the database, use the following commands:\n```bash\ncd databases/Binaries\n./extract_deploy_db.sh\n```\n\nPassword for archive is \"infected\". Warning : it contains real samples of malwares.\n\n#### Compressing database\n\nTo compress the database, use the following commands:\n```bash\n#To zip back the test database\ncd databases/Binaries\n./compress_db.sh\n```\n\n**Pypi installation and usage**\n====\n<a name=\"pypi-installation-and-usage\"></a>\n\nTo use the toolchain without docker container by using the Pypi package to install dependencies, use :\n\n```bash\npip install sema-toolchain\n```\n\nAfter cloning the git you can then use the toolchain without docker\n\n## Pypy3 usage\n\nPypy3 can be used to launch experiments, make sure to install pypy3 :\n\n```bash\nsudo add-apt-repository ppa:pypy/ppa\nsudo apt update\nsudo apt install pypy3\n```\n\nThen install the dependecies on pypy3 :\n\n```bash\npypy3 -m pip install -r /sema_scdg/requirements_pypy.txt\n```\n\n## How to use ?\n\n### Use SemaSCDG\n\nTo run experiments, run :\n```bash\npython3 sema_scdg/application/SemaSCDG.py sema_scdg/application/configs/config.ini \n```\nOr if you want to use pypy3:\n```bash\npypy3 sema_scdg/application/SemaSCDG.py sema_scdg/application/configs/config.ini \n```\n\n#### Configuration files\n\nThe parameters are put in a configuration file : `configs/config.ini`. Feel free to modify it or create new configuration files to run different experiments.\n\nThe output of the SCDG are put into `database/SCDG/runs/` by default. If you are not using volumes and want to save some runs from the container to your host machine, use :\n```bash\nmake save-scdg-runs ARGS=PATH\n```\n\n#### Parameters description\nSCDG module arguments\n\n```\nexpl_method:\n DFS Depth First Search\n BFS Breadth First Search\n CDFS Coverage Depth-First Search Strategy (Default)\n CBFS Coverage Breadth First Search\n\ngraph_output:\n gs .GS format\n json .JSON format\n EMPTY if left empty then build on all available format\n\npacking_type:\n symbion Concolic unpacking method (linux | windows [in progress])\n unipacker Emulation unpacking method (windows only)\n\nSCDG exploration techniques parameters:\n jump_it Number of iteration allowed for a symbolic loop (default : 3)\n max_in_pause_stach Number of states allowed in pause stash (default : 200)\n max_step Maximum number of steps allowed for a state (default : 50 000)\n max_end_state Number of deadended state required to stop (default : 600)\n max_simul_state Number of simultaneous states we explore with simulation manager (default : 5)\n\nBinary parameters:\n n_args Number of symbolic arguments given to the binary (default : 0)\n loop_counter_concrete How many times a loop can loop (default : 10240)\n count_block_enable Enable the count of visited blocks and instructions\n sim_file Create SimFile\n entry_addr Entry address of the binary\n\nSCDG creation parameter:\n min_size Minimum size required for a trace to be used in SCDG (default : 3)\n disjoint_union Do we merge traces or use disjoint union ? (default : merge)\n not_comp_args Do we compare arguments to add new nodes when building graph ? (default : comparison enabled)\n three_edges Do we use the three-edges strategy ? (default : False)\n not_ignore_zero Do we ignore zero when building graph ? (default : Discard zero)\n keep_inter_SCDG Keep intermediate SCDG in file (default : False)\n eval_time TODO\n\nGlobal parameter:\n concrete_target_is_local Use a local GDB server instead of using cuckoo (default : False)\n print_syscall Print the syscall found\n csv_file Name of the csv to save the experiment data\n plugin_enable Enable the plugins set to true in the config.ini file\n approximate Symbolic approximation\n is_packed Is the binary packed ? (default : False, not yet supported)\n timeout Timeout in seconds before ending extraction (default : 600)\n string_resolve Do we try to resolv references of string (default : True)\n log_level Level of log, can be INFO, DEBUG, WARNING, ERROR (default : INFO)\n family Family of the malware (default : Unknown)\n exp_dir Name of the directory to save SCDG extracted (default : Default)\n binary_path Relative path to the binary or directory (has to be in the database folder)\n fast_main Jump directly into the main function\n\nPlugins:\n plugin_env_var Enable the env_var plugin\n plugin_locale_info Enable the locale_info plugin\n plugin_resources Enable the resources plugin\n plugin_widechar Enable the widechar plugin\n plugin_registery Enable the registery plugin\n plugin_atom Enable the atom plugin\n plugin_thread Enable the thread plugin\n plugin_track_command Enable the track_command plugin\n plugin_ioc_report Enable the ioc_report plugin\n plugin_hooks Enable the hooks plugin\n```\n\nTo know the details of the angr options see [Angr documentation](https://docs.angr.io/en/latest/appendix/options.html)\n\nYou also have a script `MergeGspan.py` in `sema_scdg/application/helper` which could merge all `.gs` from a directory into only one file.\n\n#### Run multiple experiments automatically\n\nIf you wish to run multiple experiments with different configuration files, the script `multiple_experiments.sh` is available. When being in the folder containing SemaSCDG.py :\n```bash\n# To show usage\n./multiple_experiments.sh -h\n\n# Run example\n./multiple_experiments.sh -m python3 -c configs/config1.ini configs/config2.ini\n```\n\n#### Tests\n\nTo run the test :\n```bash\npython3 scdg_tests.py test_data/config_test.ini\n```\n\n#### Tutorial\n\nThere is a jupyter notebook providing a tutorial on how to use the scdg. To launch it, run\n```bash\njupyter notebook --ip=0.0.0.0 --port=5001 --no-browser --allow-root --NotebookApp.token=''\n```\nand visit `http://127.0.0.1:5001/tree` on your browser. Go to `/Tutorial` and open the jupyter notebook.\n\n### Use SemaClassifier\n\nJust run the script :\n```\npython3 SemaClassifier.py FOLDER/FILE\n\nusage: update_readme_usage.py [-h] [--threshold THRESHOLD] [--biggest_subgraph BIGGEST_SUBGRAPH] [--support SUPPORT] [--ctimeout CTIMEOUT] [--epoch EPOCH] [--sepoch SEPOCH]\n [--data_scale DATA_SCALE] [--vector_size VECTOR_SIZE] [--batch_size BATCH_SIZE] (--classification | --detection) (--wl | --inria | --dl | --gspan)\n [--bancteian] [--delf] [--FeakerStealer] [--gandcrab] [--ircbot] [--lamer] [--nitol] [--RedLineStealer] [--sfone] [--sillyp2p] [--simbot]\n [--Sodinokibi] [--sytro] [--upatre] [--wabot] [--RemcosRAT] [--verbose_classifier] [--train] [--nthread NTHREAD]\n binaries\n\nClassification module arguments\n\noptional arguments:\n -h, --help show this help message and exit\n --classification By malware family\n --detection Cleanware vs Malware\n --wl TODO\n --inria TODO\n --dl TODO\n --gspan TODOe\n\nGlobal classifiers parameters:\n --threshold THRESHOLD\n Threshold used for the classifier [0..1] (default : 0.45)\n\nGspan options:\n --biggest_subgraph BIGGEST_SUBGRAPH\n Biggest subgraph consider for Gspan (default: 5)\n --support SUPPORT Support used for the gpsan classifier [0..1] (default : 0.75)\n --ctimeout CTIMEOUT Timeout for gspan classifier (default : 3sec)\n\nDeep Learning options:\n --epoch EPOCH Only for deep learning model: number of epoch (default: 5) Always 1 for FL model\n --sepoch SEPOCH Only for deep learning model: starting epoch (default: 1)\n --data_scale DATA_SCALE\n Only for deep learning model: data scale value (default: 0.9)\n --vector_size VECTOR_SIZE\n Only for deep learning model: Size of the vector used (default: 4)\n --batch_size BATCH_SIZE\n Only for deep learning model: Batch size for the model (default: 1)\n\nMalware familly:\n --bancteian\n --delf\n --FeakerStealer\n --gandcrab\n --ircbot\n --lamer\n --nitol\n --RedLineStealer\n --sfone\n --sillyp2p\n --simbot\n --Sodinokibi\n --sytro\n --upatre\n --wabot\n --RemcosRAT\n\nGlobal parameter:\n --verbose_classifier Verbose output during train/classification (default : False)\n --train Launch training process, else classify/detect new sample with previously computed model\n --nthread NTHREAD Number of thread used (default: max)\n binaries Name of the folder containing binary'signatures to analyze (Default: output/save-SCDG/, only that for ToolChain)\n```\n\n#### Example\n\nThis will train models for input dataset\n\n```bash\npython3 SemaClassifier.py --train output/save-SCDG/\n```\n\nThis will classify input dataset based on previously computed models\n```bash\npython3 SemaClassifier.py output/test-set/\n```\n\n#### Tests\n\nTo run the classifier tests :\n```bash\npython3 classifier_tests.py configs/config_test.ini\n```\n\nCredentials\n====\n<a name=\"credentials\"></a>\n\nMain authors of the projects:\n\n* **Charles-Henry Bertrand Van Ouytsel** (UCLouvain)\n\n* **Christophe Crochet** (UCLouvain)\n\n* **Khanh Huu The Dam** (UCLouvain)\n\n* **Oreins Manon** (UCLouvain)\n\nUnder the supervision and with the support of **Fabrizio Biondi** (Avast)\n\nUnder the supervision and with the support of our professor **Axel Legay** (UCLouvain) (:heart:)\n",
"bugtrack_url": null,
"license": "BSD 2-Clause License Copyright (c) 2022, UCL-Cybersecurity team All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: 1. Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. 2. Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. ",
"summary": "Python symbolic execution package",
"version": "0.0.11",
"project_urls": {
"Homepage": "https://github.com/Manon-Oreins/SEMA-ToolChain/tree/refactor_simproc"
},
"split_keywords": [
"scdg",
" binary",
" symbolic",
" analysis"
],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "9a30bbedfe0677825e6980f01d2110ae3e7db50a8980eb2f1c7f5d214126fc86",
"md5": "c2de36b2d5442d883c70fe16e447ac0a",
"sha256": "f1e565afc33dec0f2c821c45d934d60a0e9d7a54814e5aa0c673812a10d2e7e9"
},
"downloads": -1,
"filename": "sema_toolchain-0.0.11-py3-none-any.whl",
"has_sig": false,
"md5_digest": "c2de36b2d5442d883c70fe16e447ac0a",
"packagetype": "bdist_wheel",
"python_version": "py3",
"requires_python": ">=3.8",
"size": 8152,
"upload_time": "2024-05-20T09:11:09",
"upload_time_iso_8601": "2024-05-20T09:11:09.366588Z",
"url": "https://files.pythonhosted.org/packages/9a/30/bbedfe0677825e6980f01d2110ae3e7db50a8980eb2f1c7f5d214126fc86/sema_toolchain-0.0.11-py3-none-any.whl",
"yanked": false,
"yanked_reason": null
},
{
"comment_text": "",
"digests": {
"blake2b_256": "8530031de95490470d18e522643000b04498edf89c9ffec556f534c3be599e9b",
"md5": "8af44bb80bbe4074219bd4be34f52118",
"sha256": "838ca0287e497ce7b938f1a2fafdb2bc9933141eccab3584b639aa6d6c5fd576"
},
"downloads": -1,
"filename": "sema_toolchain-0.0.11.tar.gz",
"has_sig": false,
"md5_digest": "8af44bb80bbe4074219bd4be34f52118",
"packagetype": "sdist",
"python_version": "source",
"requires_python": ">=3.8",
"size": 9455,
"upload_time": "2024-05-20T09:11:11",
"upload_time_iso_8601": "2024-05-20T09:11:11.308983Z",
"url": "https://files.pythonhosted.org/packages/85/30/031de95490470d18e522643000b04498edf89c9ffec556f534c3be599e9b/sema_toolchain-0.0.11.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2024-05-20 09:11:11",
"github": true,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"github_user": "Manon-Oreins",
"github_project": "SEMA-ToolChain",
"travis_ci": false,
"coveralls": false,
"github_actions": false,
"lcname": "sema-toolchain"
}