Skip to content

Your one-stop hub for all resources – code, documentation, and essential links – to gear up and excel in the hackathon

License

Notifications You must be signed in to change notification settings

devika-singh/resources_ai

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 

Repository files navigation

Welcome to the Multimodal Gen AI Hack, hosted by Intel Liftoff! 🚀

This hackathon focuses on Multimodal Generative AI, which enables you to create applications that combine various data types, such as text, images, and audio. You can explore two main approaches:

🔧 The Engineering-Driven Approach: Combine specialized models for each modality to create a tailored pipeline. For instance, you could use Whisper for audio transcription, pass the text to an LLM like Llama 3.1 for generating creative prompts, and then visualize the output using an image generation model like Stable Diffusion. This modular approach is well-suited for tasks such as developing an AI-driven content creation pipeline or an interactive storytelling system.

🌐 The Pure Multimodal Model Approach: Utilize state-of-the-art models like LLaVA-NeXT or Moondream, which integrate vision transformers (ViT) and LLMs into a single model. These models can process multimodal inputs end-to-end, enabling applications like visual question answering, where the system can comprehend an image and respond to queries about its content. For example, you could create an AI agent for retail that analyzes product shelf images and takes action by generating insights about restocking, suggesting promotions, or notifying relevant teams.

To make things even more exciting, we have a variety of prizes up for grabs, thanks to our fantastic startup partners in the Liftoff program.

Dive into the World of Generative AI

Dive into the world of generative AI and keep the spirit of innovation alive. Check out the challenges and resources to start your own GenAI journey today!

Generative AI

Hello, AI Adventurers! Ready to navigate the exciting world of Generative AI with us? Here's everything you need to ace the Advent of MultiModal AI hackathon:

GenAI GitHub Repositories

  1. GenAI Playground on Intel GPUs
    GitHub Repository
    A set of iPython notebooks from Stable Diffusion to LLMs.

  2. Intel AI Use Cases for GenAI
    GitHub Repository
    Key insights on GenAI applications, including Stable Diffusion, LLM inference, fine-tuning, and code generation with Intel GPUs.

  3. Diffusion Model Serving with Ray on Intel GPUs
    GitHub Repository

  4. LLM Deployments on Intel Data Center Max Series GPUs
    GitHub Repository
    A set of deployment docs for LLM deployments using TGI.

  5. LLaVA-NeXT Multimodal Chatbot with OpenVINO
    Notebook Link
    A practical example of building optimized multimodal chatbots combining vision and language with OpenVINO.


  1. Intel Developer Cloud

    • 🌐 Step into the Intel Tiber AI Cloud (IDC), register for free, and access Intel Xeons CPUs and GPUs. Discover a set of curated notebooks for Stable Diffusion, LLM inference and Finetuning on Intel under the 'Gen AI Essentials' section.
    • Each notebook will give you access to:
      • Jupyter Notebook: Each participant will work within a Jupyter Notebook environment, optimized for Generative AI challenges.
      • Disk Space: Upto 30 GB per user (Depending up on capacity).
      • GPU: Intel GPU with 48 GB (Data Center Max 1100), tailored for AI applications.
      • CPU: 4th Gen Intel Xeon.

3. Prediction Guard: Access a variety of privacy-conserving LLMs, validate outputs

  • 🛡️ Explore Prediction Guard Documentation. Check out the "Getting Started" and "Using LLMs" pages to run your first text or chat completions with the Prediction Guard API or Python client.

    import os
    import json
    import predictionguard as pg
    
    os.environ['PREDICTIONGUARD_TOKEN'] = "<your PG access token>"
    
    response = pg.Completion.create(
       model="Neural-Chat-7B",
       prompt="The advent of Gen AI hackathon is: "
    )
    
    print(json.dumps(
       response,
       sort_keys=True,
       indent=4,
       separators=(',', ': ')
    ))
  • 💪 Run through some of the examples in the Using LLMs section of the docs to learn more about basical prompting, prompt engineering, retrieval, chat, agents, etc.

  • 🌐 Dive into the Multimodal Capabilities supported by Prediction Guard APIs to build intelligent, multi-modal AI solutions.


  1. Intel Extension for PyTorch (XPUs):
    • 🔥 Check if XPU is ready:
      import torch
      import intel_extension_for_pytorch
      print(f"torch.xpu.is_available()")
    • 📚 Dive deeper at Intel XPU Tutorials

  1. Detect Your AI Resources: The Discovery Commands
    • 🔍 Uncover Intel GPUs and CPUs:
      echo "Intel GPUs:"
      xpu-smi  discovery 2> /dev/null
      echo "Intel Xeon CPU:"
      lscpu | grep "Model name"
      xpu-smi dump -m 18

  1. Python Package Installation: Effortlessly

    • 📦 Streamline your installations:
      import sys
      import site
      from pathlib import Path
      !echo "Installing..."
      !{sys.executable} -m pip cache purge > /dev/null
      !{sys.executable} -m pip install <python_package_name>
      !echo "Installation Complete."
       def get_python_version():
           return "python" + ".".join(map(str, sys.version_info[:2]))
       
       def set_local_bin_path():
           local_bin = str(Path.home() / ".local" / "bin") 
           local_site_packages = str(
               Path.home() / ".local" / "lib" / get_python_version() / "site-packages"
           )
           sys.path.append(local_bin)
           sys.path.insert(0, site.getusersitepackages())
           sys.path.insert(0, sys.path.pop(sys.path.index(local_site_packages)))
       
       set_local_bin_path()
  2. Craft Your Custom Conda Environment

    • 🧪 Mix your perfect environment:
      conda clone pytorch-gpu <new_name>
      conda activate new_name
      conda install ipykernel
      ipykernel install <>
      conda install ...
  3. Vector Databases: Exploring with LanceDB

9.Mastery in Retrieval Augmented Generation and Multi Modals

Additional Resources

Contributing 🤝

Share Your Genius:

  • Fork the repository, push your changes, and open a pull request.
  • Remember, every bit of contribution helps us sail further in the ocean of AI!

Reporting Issues 🐛

Spot Something Amiss?

  • Stumbled upon a bug or facing a challenge? Let us help! 🛠️
  • Create an issue in the GitHub repository with a detailed description.

About

Your one-stop hub for all resources – code, documentation, and essential links – to gear up and excel in the hackathon

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published