[HN Gopher] Show HN: LlamaPReview - AI GitHub PR reviewer that l...
___________________________________________________________________
Show HN: LlamaPReview - AI GitHub PR reviewer that learns your
codebase
I built LlamaPReview to solve a common frustration: most AI code
reviewers either require complex setup or don't truly understand
project context. Key differentiators: 1. One-click installation
through GitHub Marketplace - no configuration needed 2. Analyzes
your entire codebase first to understand: - Project structure -
Coding patterns - Naming conventions - Architecture decisions 3.
Completely free with no usage limits 4. Fully automated PR reviews
with zero human intervention required Technical implementation: -
Built on top of llama-github (my open source project) - Focuses on
deep code understanding rather than superficial linting - Provides
context-aware suggestions with explanations The goal is to handle
routine reviews automatically so developers can focus on complex
architectural decisions. Currently in production and processing
real PRs. Try it for free:
https://github.com/marketplace/llamapreview/ Looking for feedback
from the HN community, especially on: - What features would make
this more useful for your workflow? - How do you currently handle
code review automation? - What aspects of code understanding matter
most to you?
Author : Jet_Xu
Score : 71 points
Date : 2024-10-30 16:16 UTC (6 hours ago)
(HTM) web link (github.com)
(TXT) w3m dump (github.com)
| Roritharr wrote:
| Where's the AI Running? Where are you sending the code? Are you
| keeping some of it?
|
| I hate to be the compliance guy, but even from a startup
| perspective you'd at least want to mention what you promise to do
| here.
| KolmogorovComp wrote:
| > Where are you sending the code? Are you keeping some of it?
|
| It does not really matter for FOSS projects. For those fearing
| licence laundering, don't worry it will be done anyway for any
| public code.
| snovymgodym wrote:
| then who pays for the capacity it runs on?
| neoecos wrote:
| The delay in answering this question makes me be more careful
| about this
| Eugr wrote:
| Would be an instant no-go for any organization or individual
| that values their IP. Open Source - maybe.
| ljm wrote:
| I would want answers to all of these questions before touching
| an integration like this.
|
| The underlying library it depends on is open source, but this
| app isn't. Presumably it's holding the codebase in state.
|
| No website to speak of, just boilerplate text to satisfy
| Github's marketplace submission process.
| mistrial9 wrote:
| oh right - some one-way relationship with a corporate-or-worse
| software process that makes a record of all progress, with
| timestamps and topics.. what could go wrong?
| sprobertson wrote:
| So a git log?
| refulgentis wrote:
| Well, no, a PR contains code, not commit messages like git
| log. :)
|
| If you meant generically 'like when we store code in git', I
| believe there are some meaningful distinctions between
| voluntary version control with a host you contracted or
| built, and continuously sending code to parts unknown.
| JTyQZSnP3cQGa8B wrote:
| Yes, a changelog where the receiver says "fuck your license,
| everything's mine."
| jraph wrote:
| (Show HN)
| Squeeze2664 wrote:
| A name like llama-pr-review might help with searching for this
| thing. Preview being an actual word and all.
| nikolayasdf123 wrote:
| from your Privacy Policy, you straight up collecting users code.
| do you send it to someone else as well?
|
| might make sense for open source. closed source is no go for
| this.
| refulgentis wrote:
| Feedback:
|
| 1. Description* reeks of AI slop; it extended a surface-level
| prompt into longer surface-level insights. *: description as in
| GitHub README
|
| 2. #1 creates a situation where I go through reading this long
| thing, and realize it has no answers to even the first-level
| questions that would be on anyones mind (what model? where is it
| run?). For this to become something I'll take the time to
| integrate into my core workflow and try, it has to be *much* more
| transparent.
|
| 3. Claims in the description are ~impossible.
|
| 3b. Up front, I feel your pain, there's a hard set of constraints
| to navigate here given A) marketing needs to be concise B) people
| play fast and loose with conciseness vs. accuracy C) you need to
| sounds as good as the people in B.
|
| 3c. That being said, we're crossing into year 3 of post-ChatGPT.
| People, especially in your target audience, will know _when they
| 're reading* that you're reframing "I give text to the LLM which
| can theoratically do $X" into features, and users expect features
| to be _designed* and _intentional_. If they are, you should
| definitely highlight that to differentiate from people who just
| throw it into the LLM.
|
| 3d. Analyzes your entire repository context: impossible,
| literally, unless you're feeding it to Gemini only. I have about
| 20KLOC and its multiples of Llama context size.
|
| 3e. "Understands code relationships and dependencies" see 3c.
|
| 3f. "Contextual Analysis: Reviews code changes within the full
| repository context": see 3d.
|
| 3g. "Language Agnostic: Supports all major programming
| languages.": see 3c (is there actual work done to do this, or is
| this just "well, given I just send the text to the LLM,
| everything is supported"?)
|
| 4. nit: Should be "Show HN: LlamaPReview, AI Github PR Reviewer
| That Learns Your Codebase"
| coef2 wrote:
| I have a conundrum about this. If an LLM can learn our codebase
| and generate reasonable reviews, does this imply it could perform
| the work independently without us? Perhaps generating code and
| conducting code reviews are distinct tasks. Another related
| question is: for complex tasks that generative AI can't solve,
| could this service still provide somewhat meaningful reviews?
| Maybe it could be partially useful for certain subtasks like
| catching off-by-one errors.
| skybrian wrote:
| I'm wondering if code review is the right place to give advice;
| it seems like the process is meant for human reviews (Where there
| is latency) and pair programming might be a better metaphor for
| what AI should be doing? Earlier feedback is often better.
|
| We sort of have that with errors and warnings, where an IDE's UI
| collects them into a todo list. The trouble is, the list isn't
| necessarily prioritized very well.
|
| On the other hand, asking for a review whenever you like is easy
| to control, versus being interrupted.
|
| With all the AI tools floating around, it seems like user
| testimonials are going to be important for learning what's worth
| trying out.
| Zondartul wrote:
| By "learns" do you mean "just shove the entire codebase into the
| context window", or does actual training-on-my-data take place?
| agilob wrote:
| Description says:
|
| > Unlimited AI-powered PR reviews
|
| FAQ says:
|
| > A: Yes, we currently offer a free tier with usage limits. You
| can install and use LlamaPReview without binding any payment
| method.
|
| Only "free tier" is available.
| sksxihve wrote:
| Are people really willing to commit code that was only reviewed
| by an AI? I personally wouldn't trust that for anything that is
| customer/revenue impacting. Obvious bugs and defects aren't all
| that hard to catch in normal code reviews but subtle race
| conditions/deadlocks/memory errors can be very tricky, do you
| have examples where it shows it can catch those?
| brunooliv wrote:
| This assumes that human reviews also catch these which is
| DEFINITELY not the case either.
|
| As long as you have good pipelines, linters, a careful suite of
| tests at different levels like unit, integration, e2e and if
| you can test things in an acceptable like environment then
| human code reviews offer very very little benefit...
| sksxihve wrote:
| Your never going to catch 100% of issues, human or AI review,
| but I've found that in code reviews a lot of the benefit is
| when people ask questions about the code being reviewed and
| have a discussion on it.
|
| Is the AI tool going to ask why something was implemented in
| a way that might not match the requirement specs? Is it even
| going to know what the requirements are for the code or is it
| going to rubber stamp a review because the code looks
| reasonable?
|
| If you think human code reviews offer very very little
| benefit then you probably aren't doing them right.
| smcleod wrote:
| Hello. A few questions:
|
| - Where is the source code? This is critical for it to be
| inspected before adding to any repos. - What models are you
| using? - Where are the models running? - When you say it learns
| from your codebase is it building a RAG or similar database or
| are you fine tuning from other people's code?
| GavCo wrote:
| This reminds me of the PR Agent open source tool:
| https://github.com/Codium-ai/pr-agent
|
| I've found the code walkthroughs very useful
| lukasb wrote:
| I have a simple script I run before merging into the main branch
| that just tells Claude to look for obvious bugs, and to err on
| the side of saying it looks fine. Has stopped me from merging two
| or three bugs, 95% of the time it says things look fine so hasn't
| wasted my time.
| nbbaier wrote:
| Is that script shared somewhere?
| rplnt wrote:
| Any examples on actual PRs in public repos?
___________________________________________________________________
(page generated 2024-10-30 23:00 UTC)