Skip to content
View pillaiharish's full-sized avatar
🏠
Working from home
🏠
Working from home

Block or report pillaiharish

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
pillaiharish/README.md

πŸ‘‹ Hello, I'm Harishkumar Pillai!

Coding Animation

I'm a passionate Software Engineer with over 4 years of experience specializing in Cloud Technologies and Backend Security. My core competencies include working with Golang and Python, along with a strong background in Kubernetes, Docker, and Argo Workflow. I am exploring ways to integrate Security products with LLMs and AI Agents. I'm dedicated to designing scalable, secure infrastructure solutions and optimizing CI/CD pipelines to enhance system efficiency and reduce operational costs.


πŸš€ About Me

  • 🌟 Professional Focus: I thrive in environments where I can leverage my skills to deploy and manage cloud-based applications, particularly on platforms like AWS and Linode.
  • πŸ”§ Key Skills: Golang, Python, Kubernetes, Docker, CI/CD, Argo Workflow, GitHub Actions, Jenkins, MySQL, MongoDB, Flask.
  • πŸŽ“ Education: M.Tech in Computer Science Engineering from Institute of Technology, Nirma University.

I enjoy documenting my technical learnings and sharing insights on various tech topics. Feel free to connect with me on LinkedIn!


✍️ Latest Medium Blogs

Here are some of my latest articles where I share my insights and experiences:

Check out more on my Medium Profile!


πŸ“˜ Kubernetes Learning Blog

I'm actively learning and sharing my knowledge about Kubernetes, including its various components, best practices, and practical applications. Here are some of my recent posts:


πŸ“˜ Golang Learning Blog

I'm passionate about mastering Golang and sharing my knowledge through insightful articles. Here are some of my recent posts on Golang:


πŸ› οΈ Projects & Repositories

Feel free to explore my projects and repositories, where I apply my skills to solve real-world problems. I'm constantly learning and building, so stay tuned for more!

Github Stats


My Github Stats


πŸ’¬ Get in Touch

Let's connect and discuss all things tech! You can reach me via LinkedIn or check out my GitHub Repositories for more projects and collaborations.


Thank you for visiting my GitHub profile! πŸš€

Pinned Loading

  1. gRPC-video-streaming-server-golang gRPC-video-streaming-server-golang Public

    Video streaming using gRPC with no seeking forward or backward. The video is converted to protobuf which is served on HTTP to play on web app.

    HTML

  2. network-packet-analyser network-packet-analyser Public

    A DNS packet monitor tool which monitors the DNS domains from a list and displays the frequency of visits on a laptop.

    Go 3 2

  3. LLM-AI-Agents-Learning-Journey LLM-AI-Agents-Learning-Journey Public

    This repository is a continuously evolving collection of blogs, hands-on code, and research on Large Language Models (LLMs), AI Agents, tokenization techniques, AI security risks, and optimization …

    Jupyter Notebook 4

  4. threat-detection-analysis-tools threat-detection-analysis-tools Public

    This repo contains tools for threat detection and analysis of various malwares or any such suspicious activities.

    Go 1 3

  5. 100_models_grouped_by_vram 100_models_grouped_by_vram
    1
    |id |model_name                    |url                                                                 |type           |params_billion|typical_use                |est_vram_4bit_gb|approx_vram_gb|vram_tier |kv_cache_support                  |sharding                                                          |quant_formats                                                  |determinism_tip                                                                                                    |good_to_know                                                                                               |notes                                  |
    2
    |---|------------------------------|--------------------------------------------------------------------|---------------|--------------|---------------------------|----------------|--------------|----------|----------------------------------|------------------------------------------------------------------|---------------------------------------------------------------|-------------------------------------------------------------------------------------------------------------------|-----------------------------------------------------------------------------------------------------------|---------------------------------------|
    3
    |1  |Llama-3.1-8B-Instruct         |https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct             |LLM            |8.0           |General chat / coding      |6–7             |6.0           |4–<8 GB   |Yes (Transformer decoder; vLLM/HF)|Not needed for single-GPU                                         |GGUF (llama.cpp), AWQ, GPTQ, bitsandbytes 4/8 where available  |For repeatability: temperature 0.0–0.2, top_p=1.0, fix random seed; use greedy/beam for most deterministic outputs.|General: use 4-bit quant + paged KV cache; keep batch small on consumer GPUs.                              |128K ctx family; popular 8B.           |
    4
    |2  |Mistral-7B-v0.3               |https://huggingface.co/mistralai/Mistral-7B-v0.3                    |LLM            |7.3           |General chat               |5–6             |5.0           |4–<8 GB   |Yes (Transformer decoder; vLLM/HF)|Not needed for single-GPU                                         |GGUF (llama.cpp), AWQ, GPTQ, bitsandbytes 4/8 where available  |For repeatability: temperature 0.0–0.2, top_p=1.0, fix random seed; use greedy/beam for most deterministic outputs.|General: use 4-bit quant + paged KV cache; keep batch small on consumer GPUs.                              |Efficient 7B; many finetunes.          |
    5
    |3  |Qwen2.5-7B-Instruct           |https://huggingface.co/Qwen/Qwen2.5-7B-Instruct                     |LLM            |7.0           |General chat               |5–6             |5.0           |4–<8 GB   |Yes (Transformer decoder; vLLM/HF)|Not needed for single-GPU                                         |GGUF (llama.cpp), AWQ, GPTQ, bitsandbytes 4/8 where available  |For repeatability: temperature 0.0–0.2, top_p=1.0, fix random seed; use greedy/beam for most deterministic outputs.|General: use 4-bit quant + paged KV cache; keep batch small on consumer GPUs.                              |Long-context variants exist.           |