A simple spark standalone virtual docker compose cluster with a running thrift server for testing purposses. The cluster shares the spark warehouse space with the thrift server. Tested with Tableau.
chmod g+rwx shared-folder
sudo chown :root shared-folder
docker compose build
The final step to create your test cluster will be to run the compose file:
docker compose up -d
You should see the thrift server as an application.
docker compose run -it spark-cmd /opt/spark/bin/spark-submit --master spark://spark-master:7077 /shared-folder/load_data_write_to_server.py
docker compose run -it spark-cmd /shared-folder/connect-thrift-server.sh
docker compose up jupyter
and open the link begining with "http://127.0.0.1:8888/lab?token=".
There are some notebooks examples in the work
folder inside Jupyter.
./show-thrift-server-ip.sh