Skip to content
View Esmail-ibraheem's full-sized avatar

Block or report Esmail-ibraheem

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Esmail-ibraheem/README.md

About Me AI Research Engineer with a deep passion for neural networks, large language models, and cutting-edge machine
learning techniques. Skilled in designing and optimizing scalable systems with a strong foundation in PyTorch and CUDA. Committed to open-source development, multilingual AI accessibility, and building intuitive tools
that bridge the gap between research and real-world applications by translating them into practical implementations.

For more, check out my:

Publications: ordered by most proud of

  1. ExpertRAG: Efficient RAG with Mixture of Experts -- Optimizing Context Retrieval for Adaptive LLM Responses
  2. Galvatron: Automatic Distributed Training for Large Transformer Models
  3. Theoretical Foundations and Mitigation of Hallucination in Large Language Models
  4. Mixture of Transformers: Macro-Level Gating for Sparse Activation in Large Language Model Ensembles
  5. Bachelor Thesis: AI Engine: Deep Learning and Neural Network Engine
  6. Universal Approximation Theorem for a Single-Layer Transformer
  7. Mixture of Attention Schemes (MoAS): Learning to Route Between MHA, GQA, and MQA

Pinned Loading

  1. nanograd nanograd Public

    nanograd🧠 ML/DL and neural net ecosystem, run models like GPT, llama, stable diffusion, vision transformer, reinforcement learning, autotrainer, your Unreal Engine, but for AI, essentially making i…

    Python 10 4

  2. NeuroFlow NeuroFlow Public

    NeuroFlow is a state-of-the-art node-based platform designed to simplify the creation, management, and deployment of AI training pipelines. By combining the power of AutoTrain Advanced, a robust Re…

    Python 1

  3. Universal Universal Public

    A professional, autonomous AI assistant that works on any computer — Windows, macOS, or Linux

    Python 1

  4. Axon Axon Public

    AI research lab🔬: implementations of AI papers and theoretical research: InstructGPT, llama, transformers, diffusion models, RLHF, etc...

    Python 19 6

  5. Nexus Nexus Public

    A research-grade implementation of dynamic GPU resource allocation for Mixture of Experts models, achieving 2.3-2.4× throughput improvement and 35-45% energy efficiency gains.: https://esmail-ibrah…

    Python 6 2

  6. Transformer-pytorch Transformer-pytorch Public

    Language to Language Transformer model from scartch using pure Pytorch where I used my transformer model for translation task. from the paper "Attention is all you Need" 2017 using pytorch.

    Jupyter Notebook 9 2