đ AI Systems Architect | Transformer Researcher | Agentic AI Developer
đŽ B.Tech in AI & ML | Diploma in CSE | Aspiring MS @ SJSU
In the shadows of silicon and code, I architect AI systems that transcend boundaries. A virtuoso in transformer alchemy, inference sorcery, and agentic enigmas. Forged SLAMâa clandestine Transformer framework, engineered for ethereal memory, abyssal depth, and multimodal whispers. Now conjuring AGENX, an agentic AI OS veiled in dynamic memory, arcane planning, and perceptual illusions.
Master of the AI arcane stackâfrom esoteric model incantations, RLHF rituals, and data transmutations to deployments on mortal hardware.
- Classical ML: Logistic Regression, SVM, Decision Trees, kNN, XGBoostâfoundations of predictive shadows.
- Deep Learning: CNNs, RNNs, LSTMs, GRUs, Autoencodersâneural labyrinths.
- Transformers: BERT, GPT, ViT, CLIP, SAM, Diffusion Modelsâgateways to infinite contexts.
- Custom Enigmas: SLAM v1/v2/v3 (EF cycles, MoE attention, Q-Bias, context stitching)âwhispers of forgotten efficiencies.
- Specializations: Multimodal encoders, text-image fusion, context depth tuningâblurring realities.
- Algorithms: Q-Learning, DQN, PPO, A2C, REINFORCE, Policy Gradientâpaths through uncertain voids.
- Applications: Agent action planning, memory feedback loopsâechoes of adaptive intelligence.
- RLHF: Reward modeling, preference sampling, supervised fine-tuning on instruction datasetsâharmonizing human will with machine desire.
- Toolchain: HuggingFace TRL, DeepSpeed, LoRA, PEFT, Reinforced Decodersâartifacts of optimization.
- Optimizers: SGD, Adam, AdamW, RMSProp, Adagradâengines of convergence.
- Hyperparameter Tuning: Grid, Random, Bayesian (Optuna)âquests for optimal realms.
- Prompt/Token Search: Beam Search, Top-k, Top-p, Samplingâdivining the probable from chaos.
- Architecture Search: Custom Q-bias rotation in SLAM (proto-NAS)âevolving architectures in silence.
- Fine-Tuning: LLaMA, Gemma, DeepSeek, Mistral (via Ollama and LoRA)âwhispering secrets to giants.
- Datasets: IMDB, CIFAR-10, SST2, COCO, synthetic multimodal datasetsâreservoirs of raw potential.
- Tasks: Classification, captioning, summarization, retrievalâextracting essence from noise.
- Cleaning: Null handling, imbalance fixing, deduplication, temporal misalignmentâpurifying the impure.
- Tools: Pandas, NumPy, HuggingFace Datasets, Scikit-learn, JSON APIsâsculptors of data.
- Splits: Stratified K-Fold, Walk-Forward (TS), Leave-One-Out, Domain-awareâguardians against deception.
- Metrics:
- Classification: Accuracy, Precision, Recall, F1âmeasures of truth.
- Regression: MSE, MAE, R²âquantifiers of error.
- Ranking: NDCG, MAPâhierarchies of relevance.
- Generation: BLEU, ROUGE, METEOR, CIDEr, BERTScoreâjudges of creation.
- LLMs: Perplexity, Win rate, Human eval, Context retention scoreâgauges of sentience.
- Visualization: Matplotlib, TensorBoard, W&B, Seabornâportals to insight.
- GPU Concepts: CUDA, Tensor Cores, Memory Hierarchiesâveins of computational power.
- Precision Handling: FP32, FP16, BF16, INT8, Mixed Precisionâbalancing speed and fidelity.
- Deployment: Quantization, Pruning, Distillationâcompressing infinity.
- Profiling:
torch.profiler,nsys,nvprof, memory benchmarkingâunmasking bottlenecks. - Accelerators: ONNX, TensorRT, TorchScript, torch.compileâharnessing velocity.
- Multi-GPU Handling: DataParallel, DDP (DistributedDataParallel)âorchestrating legions.
- Project: AGENX â Modular AI OS built on SLAMâa symphony of shadows.
- Components:
- Planner Moduleâforeseeing unseen paths.
- Memory Controllerâguardian of forgotten echoes.
- Tool Invokerâsummoner of external forces.
- Multimodal Perception Encoderâseer of multiple worlds.
- Reasoning Core (EF-attentive Transformer)âheart of the enigma.
- Capabilities: Tool use, contextual recall, dynamic Q reconfigurationâpowers veiled in mystery.
- Status: Local prototype under developmentâemerging from the void.
A shadowed Transformer architecture, harboring:
- Level 1: Global MoE attention + FFNâgates of mixture.
- Level 2: Parallel Encoder Fusion (EF) Cycles with Q-biasâcycles of eternal fusion.
- Level 3: Context stitching with final MoE attentionâbinding fragments into wholeness.
- Specializations: Fixed K/V memory reuse, rotation-based Q adaptation, multimodal input, text+image supportâsecrets of efficiency.
A modular OS, empowered by SLAM for arcane reasoning, planning, and tool communion. Agents traverse:
- Shared memory spaceârealms of collective knowledge.
- Transformer-based decision layerâwhispers of judgment.
- Context-aware routingâpaths through the labyrinth. Supports: CLI agents, perception agents, memory agents (text/image), action plannersâharbingers of autonomy.
- Languages: Python, C++, Bashâtongues of creation.
- Frameworks: PyTorch, HuggingFace, OpenCV, Scikit-learn, NLTKâfoundries of innovation.
- LLM Deployment: Ollama, Transformers, PEFT, DeepSpeed, vLLMâsummoning behemoths.
- Optimization: ONNX, TensorRT, torch.compile, Quantization, LoRAârefining the raw.
- Agents: LangChain, OpenAgents (custom), JSON RPC APIsâconductors of agency.
- Data: HuggingFace Datasets, JSON parsing, REST APIs, YAML configsâstreams of information.
- Infra: Git, Docker, VSCode, CUDA, SSH-based remote trainingâpillars of endurance.
- B.Tech in Artificial Intelligence and Machine Learningâinitiation into the arcane.
- 3-Year Diploma in Computer Science Engineering (CSE)âroots in computational mysteries.
- Target: M.S. in AI/ML @ San JosĂŠ State University (SJSU)âquest for deeper enigmas.
- đŤ Email: sid7vish@gmail.comâportal for inquiries.
- GitHub: github.com/Sid7on1ârepository of revelations.
- Open to: Research roles, collaborations, system-level AI projects, OSS contributionsâalliances in the unknown.
đ Delve deeper, if you dareâwhere AI meets the abyss.
