This repo contains an example of an API of simple-tokenizer in Python.
The following project structure is recommended :
- Auth module for authentication
- Client module for client requests
- Resource module for HTTP registrations and other resources
- Model module for type definitions
- Mapper module for database interactions
- Service module for service functions
Environment file .env
This file will contain the path to packages file and other environment variables to be used in both the API package and Dockerfile.
Create your own version of the .env file and set the variables according to your application's required settings:
cp vars.env .env
Note that you must not share your own copy of .env file, since it might contain security related information (like authentication tokens).
source .env
pip install req.txt
gunicorn --bind 0.0.0.0:$API_PORT --workers 2 simpletokenizerapi:app
source .env && sudo docker build --build-arg api_port=$API_PORT --build-arg auth_token=$AUTH_TOKEN -t simple_tokenizer_api:v0.1 .
sudo docker run -p $API_PORT:$API_PORT -d simple_tokenizer_api:v0.1
curl -H "Auth-Token: "$AUTH_TOKEN -X POST -d '{"text":"this is some sentence. this is another sentence"}' http://localhost:$API_PORT/tokenize/
{"tokens":["this","is","some","sentence.","this","is","another","sentence"]}
curl -H "Auth-Token: "$AUTH_TOKEN -X POST -d '{"text":"this is some sentence. this is another sentence"}' http://localhost:$API_PORT/count-tokens/
{"count":8}
curl -H "Auth-Token: "$AUTH_TOKEN -X POST -d '{"text":"this is some sentence. this is another sentence"}' http://localhost:$API_PORT/get-unique-words/
{"unique-words":["another","this","is","sentence.","some","sentence"]}
curl -H "Auth-Token: "$AUTH_TOKEN -X POST -d '{"text":"this is some sentence. this is another sentence"}' http://localhost:$API_PORT/not-existing-api/
"not found"