
Part 1 of 2 - Part 2: Getting Confident with GitHub MCP Server: A Real QA Win
Lately I've been playing with GitHub's MCP Server to see how useful it is as a front-end developer. I wanted to write better code, faster, without constantly jumping between repos or interrupting someone to ask how an API works. What surprised me is how much it helped once I connected it to our API repos.
Suddenly I could:
- Ask questions about parts of the back-end I don't have checked out at all
- Cross-reference front-end errors with back-end logic
- Generate much nicer documentation
- Improve my types in places where I had too many
anys - Clarify assumptions before writing a PR
And all of this without cloning extra repos or relying on other developers as much.
How it helps front-end developers
Most of us have been here: a bug appears on the client, it smells like an API issue, but checking takes time or interrupts someone else's flow.
With MCP I could ask:
Can you find where this endpoint is implemented? Show me how the response is shaped.
And get a direct answer from the back-end repo. No guessing, no Slack messages, no scrolling through files I don't normally touch.
I could also ask it to compare my code against the API:
Here's the front-end function. Does it match how the API actually behaves?
It became a bridging layer between the two worlds; not replacing conversations with colleagues, just reducing the back-and-forth.
Debugging with cross-repo context
One of my favourite things was feeding MCP a front-end error and asking it to trace where the data originally came from in the back-end. It felt like having a senior engineer sitting next to me doing pair programming.
I'd give it a confusing TypeScript error, ask it to find the relevant API model, then ask it to rewrite my types to match the actual shape. It didn't always get it right, but it got me 80% of the context in seconds. That alone saved me hours.
Adding documentation to the context
Very few developers enjoy reading back-end documentation, but we all end up doing it. The big win here was that MCP could read the API repo, including the README and any existing docs, and turn it into something actionable for what I was building.
Instead of spending hours becoming an expert on a service I don't normally touch, I could ask:
Based on the README and implementation, how am I supposed to use this endpoint from the front-end?
Then follow up with questions about the response shape, required fields, and edge cases. It massively reduced the ramp-up time.
Catching AI errors before they catch you
This part is important. AI is helpful, but it's also confidently wrong sometimes. A few habits I developed that saved me from committing nonsense:
- Ask for the source of truth.
Instead of "what does this endpoint do?", ask:
Show me the exact file and code where this logic lives.
If the answer isn't what you expected, that's your sign.
- Ask it to explain its reasoning.
If something feels off:
Explain step by step how you arrived at this conclusion.
AI hallucinations usually fall apart under scrutiny.
- Use diff-style requests.
When asking for code changes:
Only show the diff, nothing else.
It forces precision and makes errors obvious.
- Keep responsibility on yourself.
The AI is fast, but you know the project best. Always double-check anything that looks too magical.
Where this leaves front-end engineering
I'm not replacing my workflow with AI. But I've added a tool that helps me ask better questions, understand the back-end faster, avoid wasting time on obvious mistakes, write cleaner types and documentation, and stay in flow without switching repos.
There are still limits; and you'll hit them quickly. But even with those limitations, the impact on my daily work has been real. I'll keep experimenting, especially with automating PR reviews and generating tests, and I'll share more once I've broken things enough times.
Written by Tania Papazafeiropoulou - Senior Front-End Engineer & Tech Lead.