Skip to content

Create AI Powered Workflows to scrape the Web πŸ”βš™οΈπŸ”πŸ› οΈ

License

Notifications You must be signed in to change notification settings

BernieTv/Workflows-App

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

16 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

πŸš€ AI-Powered Visual Web Scraper Builder πŸ”§

This Full Stack SaaS application allows users to visually build, manage, and schedule web scrapers using a workflow builder powered by AI. πŸ€– Users can create, modify, and delete workflows with an intuitive drag-and-drop interface. The integration of AI simplifies the web scraping process, making it accessible to both technical and non-technical users. 🌐

πŸ”₯ Features

🌟 Key Features

  • πŸ” Visual Workflow Builder: Drag-and-drop interface to design scraping workflows effortlessly. πŸ”§
  • πŸ€– AI Assistance: AI-powered suggestions for selectors, workflow optimization, and error handling. 🌐
  • πŸ”‘ Credential Management: Securely manage login credentials for scraping protected websites. πŸ”’
  • ⏳ Scheduling System: Set up automatic scraping schedules for periodic data extraction. ⏰
  • πŸ›οΈ Workflow Management: Create, modify, delete, and duplicate workflows with ease. πŸ”„
  • πŸ“„ Data Export: Export scraped data in various formats (e.g., CSV, JSON). πŸ“Š

⚑ Built with Next.js

  • 🌐 Server-Side Rendering (SSR) for optimized SEO and performance.
  • 🌍 API Routes to handle backend logic.
  • πŸ“ Dynamic Routing for user-specific workflows.
  • πŸ” Built-in Authentication using Clerk Authentication for secure user sessions.

βš™οΈ Prerequisites

  • πŸ’» Node.js (v16 or later)
  • πŸ“Š PostgreSQL database
  • πŸ”§ API Key for OpenAI

πŸ”Ž Steps

  1. πŸ”„ Clone the repository

    git clone https://github.com/BernieTv/Workflows-App.git
    cd Workflows-App
  2. πŸ“¦ Install dependencies

    npm install
  3. πŸ” Set up environment variables Create a .env file in the root directory and add the following:

    NEXT_PUBLIC_OPENAI_API_KEY=your_openai_api_key
    DATABASE_URL=postgresql://username:password@localhost:5432/yourdb
    NEXT_PUBLIC_CLERK_PUBLISHABLE_KEY=your_clerk_secret
    CLERK_SECRET_KEY=sk_
  4. πŸ”§ Run database migrations

    npx prisma migrate dev
  5. πŸš€ Start the development server

    npm run dev

    The app will be available at http://localhost:3000. 🏠


πŸ”§ Usage

1. πŸ” Sign Up/Log In

  • Use Clerk Authentication to sign up or log in to your account. πŸ”‘

2. πŸ”§ Create a Workflow

  • Drag and drop nodes to define scraping tasks. 🌐
  • Use AI suggestions for selector optimization. πŸ€–

3. πŸ”‘ Set Credentials

  • Securely store website login credentials if required. πŸ”’

4. ⏳ Schedule Scraping

  • Use the scheduling feature to automate scraping tasks. πŸ•°οΈ

5. πŸ“„ Export Data

  • Download scraped data in the desired format. πŸ“Š

πŸ’ͺ Development

βš™οΈ Scripts

  • πŸš€ Start development server: npm run dev
  • 🌍 Build for production: npm run build
  • 🏠 Run production server: npm start

πŸ”§ Linting and Formatting

  • πŸ”’ Lint code: npm run lint
  • πŸ”„ Format code: npm run format

About

Create AI Powered Workflows to scrape the Web πŸ”βš™οΈπŸ”πŸ› οΈ

Topics

Resources

License

Stars

Watchers

Forks

Languages