Terality Docs
  • What is Terality?
  • Documentation
    • Quickstart
      • Setup
      • Tutorial
      • Next steps
    • User guide
      • Supported configurations
      • User dashboard
      • Importing and exporting data
        • Reading/writing to storage
        • Reading from multiple files
        • Writing to multiple files
        • Storage services
        • Data formats
        • From/to pandas
      • .apply() and passing callables
      • Caching
      • Best practices and anti-patterns
      • Upgrading your client version
      • Client configuration (CLI)
      • Support for external packages
      • Advanced topics
        • Data mutability and ownership: Terality vs pandas
    • API Reference
      • Conversion from/to pandas
      • Write to multiple files
    • Deploy Terality in your own AWS account
    • Releases
  • FAQ
    • Differences with
      • Terality and Pandas
      • Terality vs Spark
      • Terality vs Dask
    • Pricing
    • Security
    • Support & contact
    • Common setup issues
Powered by GitBook
On this page
  • Option 1: download and run the tutorial on your computer
  • Option 2: view the tutorial on GitHub

Was this helpful?

  1. Documentation
  2. Quickstart

Tutorial

Start using Terality on a big dataset that we provide for you.

PreviousSetupNextNext steps

Last updated 3 years ago

Was this helpful?

Let's process some data with Terality!

This tutorial explores a 5 GB dataset from Reddit, and can easily run on the Terality free plan (of course, Terality works fine with much bigger datasets).

Option 1: download and run the tutorial on your computer

If you completed the Terality setup, we recommend downloading the tutorial and running it yourself as you go.

Download the Jupyter notebook from this link:

Option 2: view the tutorial on GitHub

View the tutorial on GitHub .

here
https://terality-public.s3.amazonaws.com/tutorial/Tutorial_5GB.ipynb
Jupyter notebook on 5GB dataset