A micro-CRM to help Code For Denver discover leads and manage its outreach to potential partners.
- Socrata API
- Dataset with registered business entities in Colorado. It can be filtered to return only nonprofits.
- Colorado Nonprofit Association
- Website with nonprofit members registered with Colorado Nonprofit Association.
- Twitter?
- LinkedIn?
- Go to the project's github page.
- Find the green "Code" button
- Click the clipboard icon to copy a link to the git repo.
- In a terminal, navigate to the directory where you want to create the project folder and clone the repo:
git clone <git-repo-name>
- Install Docker and Docker-compose
- Install node
- Install node dependencies:
cd
tofrontend/cfd-partner-finder-frontend
npm i --save
cd ../..
- Run the frontend, rest api, and database in docker containers:
docker-compose up --build -d
- Check the containers are running
docker ps
- Try connecting to the database with psql:
to exit psql, type
docker exec -it partner-finder_postgres_1 psql -U cfd_partner_finder select * from leads limit 5;
\q
- Check that the api works with curl:
- try the healthcheck endpoint:
curl http://localhost:8000/healthcheck
- get an access token to use other api endpoints:
curl --location --request POST 'http://localhost:8000/login' \ --header 'Content-Type: application/json' \ --data-raw '{ "username": "admin", "password": "password" }'
- get a list of leads:
curl --location --request GET 'http://localhost:8000/leads' \ --header 'Authorization: Bearer <insert your token here>'
- try the healthcheck endpoint:
- Check that the frontend is working:
- In a browser, go to http://localhost:3000
- You should see a login page. Use these credentials to continue to the homepage:
- username:
[email protected]
- password:
password
- username:
You'll need a python virtual environment in the backend
directory. Make sure you have python 3.7 or up installed. Ideally 3.9 since that is what is used in the rest api. You can check the version with python --version
Change into the backend directory then do python -m venv venv
. This should create a venv
directory.
Next you'll want to activate the virtual environment with source venv/bin/activate
.
Then install requirements with pip install -r requirements.txt
You should also need to set some environment variables so alembic can send queries to the locally running database. Create a .env
file with touch .env
, then add these lines to it:
export FLASK_APP=api/app:dev_app
export FLASK_ENV=development
export POSTGRES_PASSWORD=password
export POSTGRES_USER=cfd_partner_finder
export POSTGRES_DB=cfd_partner_finder
export POSTGRES_HOST=localhost
export POSTGRES_PORT=5432
export ALLOW_CORS=true
export SECRET_KEY=supersafe
export PYTHONPATH="${pwd}"
Now source the environment variables: source .env
Finally, you can create a new migration by doing alembic revision -m "<description of migration>"
. This should create a new file under the versions
directory.
The backend generates swagger documentation. This is a webpage that lets you make interactive api calls to test out the rest api before using it in your code. To run the swagger docs locally, make sure the api
docker service is running. Check the api logs for a bearer token that you can use to authenticate on the swagger page. If you ran docker compose with the -d
flag, you can get the logs with docker compose logs api
.
Now look for bearer tokens that let you authenticate as a normal user and as an admin:
api_1 | [2021-08-11 01:05:47 +0000] [19] [INFO] To authenticate as [email protected], include this header with the request:
api_1 | Authorization: Bearer eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJ1c2VybmFtZSI6InVzZXJAZ21haWwuY29tIiwiZXhwaXJlcyI6IjIwMjEtMDgtMTJUMDE6MDU6NDcuNTM0NDgzKzAwOjAwIiwiYWRtaW4iOmZhbHNlfQ.41xKVHDz0ONRiWx-fWqifVvDBSzCN6vPmmf4ZWV0H3g
api_1 | [2021-08-11 01:05:47 +0000] [19] [INFO] To authenticate as [email protected], include this header with the request:
api_1 | Authorization: Bearer eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJ1c2VybmFtZSI6ImFkbWluQGdtYWlsLmNvbSIsImV4cGlyZXMiOiIyMDIxLTA4LTEyVDAxOjA1OjQ3LjU2NTExNCswMDowMCIsImFkbWluIjp0cnVlfQ.NNUMN92roOU44DKXcnstBUK_vpRfg57RYJyBMCuSdmQ
Copy only the value of the header, that is, the part that looks like
Bearer eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJ1c2VybmFtZSI6InVzZXJAZ21haWwuY29tIiwiZXhwaXJlcyI6IjIwMjEtMDgtMTJUMDE6MDU6NDcuNTM0NDgzKzAwOjAwIiwiYWRtaW4iOmZhbHNlfQ.41xKVHDz0ONRiWx-fWqifVvDBSzCN6vPmmf4ZWV0H3g
Now in a web browser, navigate to http://localhost:8000/apidocs . You should see a page that looks like this:
Click on the green "Authorize" button with the lock icon, paste the bearer token you copied into the login form, and click "Authorize".
To send a request to any of the endpoints, click on one of the colored boxes, then click "Try it out" in the upper right corner. This lets you edit the request parameters and body through a form. You can send the request and view the response with the "Execute" button.
We have github actions that will check that backend code is in the correct format and abides by PEP8 standards. You will need to run a formatter and a linter on your code before committing in order for your changes to be accepted. In the backend/scripts
, directory, there are scripts called lint.sh
and format.sh
for doing this. You can run them directly from the backend
directory:
cd backend
source venv/bin/activate
chmod +x scripts/*.sh
./scripts/format.sh
./scripts/lint.sh
After running lint.sh
, you should see an output of 0
if everything is okay. Otherwise flake8 will output lines that need to be changed.
Once you've made formatting and linting changes, make a commit with a message like lint and format
and add it to your PR. It is helpful to PR reviewers if you keep your formatting changes in their own commit because they can potentially make it harder to read your other code changes.
Postman is a web client for testing out REST apis. See here to view and export postman requests for this project. You will also need to install postman, import the collection, and then run the api on localhost to use postman in development.
- Make sure python 3 is installed on your system
- from the project root directory, change to the data analysis directory
cd ./data_analysis
- Create a virtual environment
python3 -m venv --prompt data_analysis venv
- You should see a newly created folder called
venv
- You should see a newly created folder called
- Activate the virtual environment
source venv/bin/activate
- Your terminal prompt should change to display
(data_analysis)
on the left while the virtual environment is active.
- Your terminal prompt should change to display
- Upgrade the virtual environment's installation of pip
pip install --upgrade pip
- Install dependencies
pip install -r requirements.txt
- Run a jupyter server:
jupyter notebook
- You should see a file system open in a web browser. If not, go to http://localhost:8888/tree
- Click on
notebooks
, and thenbusinesses.ipynb
. You should now see a notebook - When you are done, stop the jupyter server with
Ctrl+C
and deactivate the virtual environment withdeactivate
.