arghwhat 20 hours ago

> no way of running actions locally

My policy is to never let pipeline DSLs contain any actual logic outside orchestration for the task, relying solely on one-liner build or test commands. If the task is more complicated than a one-liner, make a script for it in the repo to make it a one-liner. Doesn't matter if it's GitHub Actions, Jenkins, Azure DevOps (which has super cursed yaml), etc.

This in turn means that you can do what the pipeline does with a one-liner too, whether manually, from a vscode launch command, a git hook, etc.

This same approach can fix the mess of path-specific validation too - write a regular script (shell, python, JS, whatever you fancy) that checks what has changed and calls the appropriate validation script. The GitHub action is only used to run the script on PR and to prepare the CI container for whatever the script needs, and the same pipeline will always run.

  • Tainnor 19 hours ago

    The reason why many CI configs devolve into such a mess isn't typically that they don't extract complicated logic into scripts, it's about all the interactions with the CI system itself. This includes caching, sharing of artifacts, generating reports, configuring permissions, ordering of jobs, deciding when which jobs will run, deciding what to do when jobs fail, etc. All of this can get quite messy in a large enough project.

    • bluGill 4 hours ago

      You should generate your report with regular scripts. You need ci config to deploy them but that is the only part that should be different.

      • evilduck 2 hours ago

        This doesn’t really work when you start sharding tests.

    • pydry 8 hours ago

      It never becomes unbearably messy this way though.

      The reason it gets unbearably messy is because most people google "how to do x in github actions" (e.g. send a slack message) and there is a way and it's almost always worse than scripting it yourself.

  • 0xbadcafebee 6 hours ago

    100%. The ci/cd job should be nothing more than a wrapper around the actual logic which is code in your repo.

    I write a script called `deploy.sh` which is my wrapper for my ci/cd jobs. It takes options and uses those options to find the piece of code to run.

    The ci/cd job can be parameterized or matrixed. The eventually-run individual jobs have arguments, and those are passed to deploy.sh. Secrets/environment variables are set from the ci/cd system, also parameterized/matrixed (or alternately, a self-hosted runner can provide deploy.sh access to a vault).

    End result: from my laptop I can run `deploy.sh deploy --env test --modules webserver` to deploy the webserver to test, and the CI/CD job also runs the same job the same way. The only thing I maintain that's CI/CD-specific is the GitHub Action-specific logic of how to get ready to run `deploy.sh`, which I write once and never change. Thus I could use 20 different CI/CD systems, but never have to refactor my actual deployment code, which also always works on my laptop. Vendor lock-in is impossible, thanks to a little abstraction.

    (If you have ever worked with a team with 1,000 Jenkins jobs and the team has basically decided they can never move off of Jenkins because it would take too much work to rewrite all the jobs, you'll understand why I do it this way)

    • hardwaresofton 4 hours ago

      Hey if you’ve never heard of it consider using just[0], it’s a better makefile and supports shell scripting explicitly (so at least equivalent in power, though so is Make)

      [0]: https://github.com/casey/just

      • 0xbadcafebee 3 hours ago

        Thank you, I have seen it, but I prefer Make.

  • hinkley 8 hours ago

    I’ll go so far as to say the massive add on/plugin list and featuritis of CI/CD tools is actively harmful to the sanity of your team.

    The only functionality a CI tool should be providing is:

    - starting and running an environment to build shit in

    - accurately tracking success or failure

    - accurate association of builds with artifacts

    - telemetry (either their own or integration) and audit trails

    - correlation with project planning software

    - scheduled builds

    - build chaining

    That’s a lot, but it’s a lot less than any CI tool made in the last 15 years does, and that’s enough.

    There’s a big difference for instance between having a tool that understands Maven information enough to present a build summary, and one with a Maven fetch/push task. The latter is a black box you can’t test locally, and your lead devs can’t either, so when it breaks, it triggers helplessness.

    If the only answer to a build failure is to stare at config and wait for enlightenment, you fucked up.

  • ironfootnz 14 minutes ago

    That’s my policy too. I see way too many Jenkins/Actions scripts with big logic blocks jammed into YAML. If the entire build and test process is just a single script call, we can run it locally, in a GitHub workflow, or anywhere else. Makes it less painful to switch CI systems, and devs can debug easily without pushing blind commits. It’s surprising how many teams don’t realize local testing alone saves huge amounts of time.

  • ukoki 7 hours ago

    > My policy is to never let pipeline DSLs contain any actual logic outside orchestration for the task,

    I call this “isomorphic CI” — ie: as long as you set the correct env vars, it should run identically on GitHub actions, Jenkins, your local machine, a VM etc

    • reactordev 7 hours ago

      This is the only DevOps way. Abstract the build into a single step.

      • vvillena 6 hours ago

        And yet, you would be surprised at the amount of people who react like that's an ignorant statement ("not feasible in real world conditions"), an utopic goal ("too much time to implement"), an impossible feat ("automation difficults human oversight"), or, my favorite, the "this is beneath us" excuse ("see, we are special and this wouldn't work here").

        Automation renders knowledge into a set of executable steps, which is much better than rendering knowledge into documentation, or leaving it to rot in people's minds. Compiling all rendered knowledge into a single step is the easiest way to ensure all elements around the build and deployment lifecycle work in unison and are guarded around failures.

  • cnunciato 3 hours ago

    >Lack of local development. It's a known thing that there is no way of running GitHub Actions locally.

    This is one thing I really love about Buildkite[0] -- being able to run the agent locally. (Full disclosure: I also work for Buildkite.) The Buildkite agent runs as a normal process too (rather than as a Docker container), which makes the process of workflow development way simpler, IMO. I also keep a handful of agents running locally on my laptop for personal projects, which is nice. (Why run those processes on someone else's infra if I don't have to?)

    >Reusability and YAML

    This is another one that I believe is unique to Buildkite and that I also find super useful. You can write your workflows in YAML of course -- but you can also write them in your language of choice, and then serialize to YAML or JSON when you're ready to start the run (or even add onto the run as it's running if you need to). This lets you encapsulate and reuse (and test, etc.) workflow logic as you need. We have many large customers that do some amazing things with this capability.

    [0]: https://buildkite.com/docs/agent/v3

  • jicea 18 hours ago

    I don't understand why this is not the evident approach for everyone writing GitHub Actions/GitLab CI/CD yaml etc....

    I've struggled in some teams to explained why it's better to extract your command in scripts (ShellCheck on it, scripts are simple to run locally etc...) instead of writing a Frankenstein of YAML and shell commands. I hope someday to find an authoritative guidelines on writing pipeline that promote this approach so at least I can point to this link instead of defending myself being a dinosaur!

    • mrweasel 18 hours ago

      In a previous job we had a team tasked with designing these "modern" CI/CD pipeline solutions, mostly meant for Kubernetes, but it was suppose to work for everything. They had such a hard on for tools that would run each step as a separate isolated task and did not want pipelines to "devolve" into shell scripts.

      Getting anything done in such environments are just a pain. You spend more time fighting the systems than you do actually solving problems. It is my opinion that a CI/CD system needs just the following features: Triggers (source code repo, http endpoints or manually triggered), secret management and shell script execution. That's it, you can build anything using that.

      • eddd-ddde 15 hours ago

        I think what they really wanted was something like bazel. The only real benefit I can think right now for not "devolving" into shell scripts is distributed caching with hermetic builds. It has very real benefits but it also requires real effort to work correctly.

      • chrisweekly 4 hours ago

        Mostly agreed, but (maybe orthogonal) IME, popular CI/CD vendors like TeamCity* can make even basic things like shell script execution problematic.

        * TC offers sh, full stop. If you want to script something that depends on bash, it's a PITA and you end up with a kludge to run bash in sh in docker in docker.

        • mdaniel an hour ago

          > If you need a specific interpreter to be used, specify shebang (for example, #!/bin/bash) as the first line of the script.

          https://www.jetbrains.com/help/teamcity/command-line.html#:~...

          Your "docker in docker" comment makes me wonder if you're conflating the image that you are using, that just happens to be run by TeamCity, versus some inherent limitation of TC. I believe a boatload of the Hashicorp images ship with dash, or busybox or worse, and practically anything named "slim" or "alpine" similarly

      • datavirtue 6 hours ago

        I just joined as the enterprise architect for company that has never had one. There is an existing devops team that is making everyone pull their hair out and I haven't had a single spare minute to dig in on their mess but this sounds early familiar.

        • mongol 5 hours ago

          Is this really the job of an enterprise architect? To dig into the details of a devops team's mess?

          • datavirtue 3 hours ago

            It's more about giving the team and the overall strategy a thumbs up or down, so yes.

    • 0xbadcafebee 2 hours ago

      I think I can summarize it in a rough, general way.

        CI/CD is a method to automate tasks in the background that you would otherwise run on your laptop. The output
        of the tasks are used as quality gates for merging commits, and for deployments.
        
        - Step 1. Your "laptop in the cloud" requires some configuration (credentials, installed software, cached artifacts)
          before a job can be run.
          - Requires logic specific to the CI/CD system
        
        - Step 2. Running many jobs in parallel, passing data from step to step, etc requires some instructions.
          - Requires logic specific to the CI/CD system
        
        - Step 3. The job itself is the execution of a program (or programs), with some inputs and outputs.
          - Works the same on any computer (assuming the same software, environment, inputs, etc)
          - Using a container in Step 1. makes this practical and easy
        
        - Step 4. After the job finishes, artifacts need to be saved, results collected, and notifications sent.
          - Some steps are specific to the CI/CD system, others can be a reusable job
        
        Step 3 does not require being hard-coded into the config format of the CI/CD system. If it is instead
        just executable code in the repo, it allows developers to use (and work on) the code locally without
        the CI/CD system being involved. It also allows moving to a different CI/CD system without ever rewriting
        all the jobs; the only thing that needs to be rewritten are the CI/CD-specific parts, which should be
        generic and apply to all jobs pretty much the same.
        
        Moving the CI/CD-specific parts to a central library of configuration allows you to write some code
        once and reuse it many times (making it DRY). CircleCI Orbs, GitHub Actions, Jenkins Shared Libraries/
        Groovy Libraries, etc are examples of these. Write your code once, fix a bug once, reuse it everywhere.
    • arghwhat 18 hours ago

      My "favorite" is when I see people go all in, writing thousands of lines of Jenkins-flavor Groovy that parses JSON build specifications of arbitrary complexity to sort out how to build that particular project.

      "But then we can reuse the same pipeline for all our projects!"

      • ozim 6 hours ago

        I think that is pitfall of software devs.

        For me it was an epiphany as software dev - not to write reusable extensible scripts - I am so much more productive after that.

      • baby_souffle 2 hours ago

        > "But then we can reuse the same pipeline for all our projects!"

        oh god just reading that gave me PTSD flash backs.

        At $priorGig there was the "omni-chart". It was a helm chart that was so complex it needed to be wrapped in terraform and used composable terraform modules w/ user var overrides as needed.

        Debugging anything about it meant clearing your calendar for the day and probably the following day, too.

      • lowercased 7 hours ago

        I can rarely reuse the same pipeline for the same project 6 months down the road, much less reuse for anything else.

        The few bits that end up getting reused are the externalized bash scripts.

    • maccard 18 hours ago

      To make the thing actually fast at scale, a lot of the logic ends up being specific to the provider; requiring tokens, artifacts etc that aren't available locally. You end up with something that tries to detect if you're running locally or in CI, and then you end up in exactly the same situation.

    • arccy 7 hours ago

      it can be quite hard to write proper scripts that work consistently... different shells have different behaviours, availability of local tools, paths, etc

      and it feels like fighting against the flow when you're trying to make it reusable across many repos

      • akdev1l 7 hours ago

        Containerize the build environment so everything is captured (dependencies, build tools, etc)

        • carlmr 6 hours ago

          If you're not containerizing your CI/CD, you're really lost.

          • lmm 7 minutes ago

            Only if your tech stack is bad (i.e. Python). My maven builds work anywhere with an even vaguely recent maven and JVM (and will fail-fast with a clear and simple error if you try to run them in something too old), no need to put an extra layer of wrapping around that.

          • maccard 5 hours ago

            How do I containerize building desktop apps for windows with MSVC?

          • akoboldfrying 2 hours ago

            That might be the case if Docker did in fact guarantee (or at least make it easy to guarantee) deterministic builds -- but it doesn't really even try:

            1. Image tags ("latest", etc.) can change over time. Does any layer in your Dockerfile -- including inside transitive deps -- build on an existing layer identified by tag? If so, you never had reproducibility.

            2. Plenty of Dockerfiles include things like "apt-get some-tool" or its moral equivalent, which will pull down whatever is the latest version of that tool.

            It's currently common and considered normal to use these "features". Until that changes, Docker mostly adds only the impression of reproducibility, but genuine weight and pain.

        • darthwalsh 6 hours ago

          I'm not sold on using containers for macOS desktop apps...

    • dboreham 4 hours ago

      The reason for this is that nobody took the time to write a proper background document on Github Actions. The kind of information that you or I might convey if asked to explain it at the whiteboard to junior hires, or senior management.

      This syndrome is very common these days. Things are explained differentially: it's like Circle CI but in the GH repo. Well that's no use if the audience wasn't around when Circle CI was first new and readily explained (It's like Jenkins but in the cloud...).

  • rtpg 5 hours ago

    The actual subtle issue here is that sometimes you actually need CI features around caching and the like, so you are forced to engage with the format a bit.

    You can, of course, chew it down to a bare minimum. But I really wish more CI systems would just show up with "you configure us with scripts" instead of the "declarative" nonsense.

    • Guvante 5 hours ago

      CI that isn't running on your servers wants very deep understanding of how your process works so they can minimize their costs (this is true whether or not you pay for using CI)

      • rtpg 4 hours ago

        Totally! It's a legitimate thing! I just wish that I had more tools for dynamically providing this information to CI so that it could work better but I could also write relatively general tooling with a general purpose language.

        The ideal for me is (this is very silly and glib and a total category error) LSP but for CI. Tooling that is relatively normalized, letting me (for example) have a pytest plugin that "does sharding" cleanly across multiple CI operators.

        There's some stuff and conventions already of course, but in particular caching and spinning up jobs dynamically are still not there.

  • alkonaut 18 hours ago

    That’s usually very hard or impossible for many things. The AzDo yaml consists of a lot of steps that are specific to the CI environment (fetching secrets, running tests on multiple nodes, storing artifacts of various kinds).

    Even if the ”meat” of the script is a single build.ps oneliner, I quickly end up with 200 line yaml scripts which have no chance of working locally.

    • arghwhat 17 hours ago

      Azure DevOps specifically has a very broken approach to YAML pipelines, because they effectively took their old graphical pipeline builder and just made a YAML representation of it.

      The trick to working with this is that you don't need any of their custom Azure DevOps task types, and can use the shell type (which has a convenient shorthand) just as well as in any other CI environment. Even the installer tasks are redundant - in other CI systems, you either use a container image with what you need, or install stuff at the start, and Azure DevOps works with both of these strategies.

      So no, it's neither hard nor impossible, but Microsoft's half-assed approach to maintaining Azure DevOps and overall overcomplicated legacy design makes it a bit hard to realize that doing what their documentation suggests is a bad idea, and that you can use it in a modern way just fine. At least their docs do not recommend that you use the dedicated NPM-type task for `npm install` anymore...

      (I could rant for ages about Azure DevOps and how broken and unloved it is from Microsoft's side. From what I can tell, they're just putting in the minimum effort to keep old Enterprise customers that have been there through every rename since Team Foundation Server from jumping ship - maybe just until Github's enterprise side has matured enough? Azure DevOps doesn't even integrate well with Azure, despite its name!)

      • mplanchard an hour ago

        This doesn’t seem to address the parent comment’s point at all, which was about required non-shell configuration such as for secrets, build parallelism, etc.

  • mhh__ 3 hours ago

    Indeed. Anything else is just asking for trouble.

    CI must only be an application of tools available to developers locally.

    This, other than for practical reasons, is a tax on complexity.

  • sirlone an hour ago

    So for example, my action builds on 4 different platforms (win-64, linux-amd64, mac-intel, mac-arm), it does this in parallel then gets the artifacts for all for and bundles them into a single package.

    How would you suggest I do this following your advice?

  • riperoni 16 hours ago

    I agree with wrapping things like build scripts to test locally.

    Still, some actions or CI steps are also not meant to be run locally. Like when it publishes to a repo or needs any credentials that are used by more than one person.

    Btw, Github actions and corresponding YAML are derived from Azure DevOps and are just as cursed.

    The whole concept of Github CI is just pure misuse of containers when you need huge VM images - container is technically correct, but a far fetched word for this - that have all kinds of preinstalled garbage to run typescript-wrapped code to call shell scripts.

  • InvertedRhodium 7 hours ago

    I use gitlab-ci-local to run Gitlab pipelines locally - does such a thing not exist for GitHub actions?

    • mdaniel 6 hours ago

      Oh, yeah, I remember looking at that a while back. I don't recall how much it had implemented at the time but it seems that firecow took a vastly different approach than nektos/act did, going so far as to spend what must have been an enormous amount of time/energy to cook up https://github.com/firecow/gitlab-ci-local/blob/4.56.2/src/s... (and, of course, choosing a dynamically typed language versus golang)

  • thatxliner 4 hours ago

    How would you set up tool installations? Inside the CI or inside the script?

  • nunez 14 hours ago

    While youre correct, environmental considerations are another advantage that testing locally SHOULD be able to provide (i.e. you can test your scripts or Make targets or whatever in the same runner that runs in the actual build system.)

    This is not possible with GHA.

    • arghwhat 14 hours ago

      Of course you can, just specify a container image of your choice and run the same container for testing locally.

      However, replicating environmental details is only relevant where the details are known to matter. A lot of effort has been wasted and workflows crippled by the idea that everything must be 100% identical irrespective of actual dependencies and real effects.

  • potamic 17 hours ago

    This is the right way to use CI/CD systems, as dumb orchestrators without inherent knowledge of your software stack. But the problem is, everything from their documentation, templates, marketplace encourage you to do exactly the opposite and couple your build tightly with their system. It's poor product design imo, clearly optimising for vendor lock-in over usability.

  • no_wizard 7 hours ago

    Folks pick the wrong tool for the job at hand.

    I suspect the author of the article could greatly simplify matters if they used a task running tool to orchestrate running tasks, for example. Pick whatever manner of decoupling you want really, most of the time this is the path to simplified CI actions. CI is best when its thought of as a way to stand up fresh copies of an environment to run things inside of.

    I have never had the struggles that so many have had with CI as a result. Frankly, I'm consistently surprised at how overly complex people make their CI configurations. There's better tools for orchestration and dependency dependent builds, which is not its purpose to begin with.

    • ledauphin 6 hours ago

      I generally agree with you, but I'd be interested to hear your take on what the purpose of CI _actually is_.

      It seems to me that a big part of the problem here (which I have also seen/experienced) is that there's no one specific thing that something like GitHub Actions is uniquely suited for. Instead, people want "a bunch of stuff to happen" when somebody pushes a commit, and they imagine that the best way to trigger all of that is to have an incredibly complex - and also bespoke - system on the other end that does all of it.

      It's like we learned the importance of modularity in the the realm of software design, but never applied what we learned to the tools that we work with.

      • no_wizard 2 hours ago

        Standing up fresh images for validation and post validation tasks.

        CI shines for running tests against a clean environment for example.

        Really any task that benefits from a clean image being stood before running a task.

        The key though is to decouple the tasks from the CI. Complexity like pre-packaging artifacts is not a great fit for CI configuration, that is best pushed to a tool that doesn’t require waterfall logic to make it work.

        There is a reason task runners are very popular still

  • izietto 19 hours ago

    This. I'd go more relaxed on the one-liner requirement, a few lines are fine, but the approach is correct IMHO.

ironfootnz 15 minutes ago

I’ve seen many teams get stuck when they rely too heavily on GitHub Actions’ magic. The key issue is how tightly your build logic and config become tied to one CI tool. If the declarative YAML gets too big and tries to handle complex branching or monorepos, it devolves into a maintenance headache—especially when you can’t test it locally and must push blind changes just to see what happens.

A healthier workflow is to keep all the logic (build, test, deploy) in portable scripts and let the CI only orchestrate each script as a single step. It’s easier to troubleshoot, possible to run everything on a dev machine, and simpler if you ever migrate away from GitHub.

For monorepos, required checks are maddening. This should be a first-class feature where CI can dynamically mark which checks apply on a PR, then require only those. Otherwise, you do hacky “no-op” jobs or you force your entire pipeline to run every time.

In short, GitHub Actions can be powerful for smaller codebases or straightforward pipelines, but if your repo is big and you want advanced control, it starts to feel like you’re fighting the tool. If there’s no sign that GitHub wants to address these issues, it’s totally reasonable to look elsewhere or build your own thin orchestration on top of more flexible CI runners.

benrutter 19 hours ago

Oh boy, there's a special kind of hell I enter into everytime I set up new github actions. I wrote a blog post a few months ago about my pain[0] but one of the main things I've found over the years is you can massively reduce how horrible writing github actions is by avoiding prebuilt actions, and just using it as a handy shell runner.

If you write behaviour in python/ruby/bash/hell-rust-if-you-really-want and leave your github action at `run: python some/script.py` then you'll have something that's much easy to test locally, and save yourself a lot of pain, even if you wind up with slightly more boilerplate.

[0] https://benrutter.github.io/posts/github-actions/

  • jbaber an hour ago

    I had this idea the other day when dealing with CI and thought it must be dumb because everyone's not already doing it. It would make your CI portable to other runners in future, too.

  • riperoni 16 hours ago

    At this point, just pause with Github Actions and compare it to how GiLab handles CI.

    Much more intuitive, taking shell scripts and other script commands natively and not devolving into a mess of obfuscated typescript wrapped actions that need a shit ton of dependencies.

    • Aeolun 6 hours ago

      The problem with Gitlab CI is that now you need to use Gitlab.

      I’m not even sure when I started feeling like that was a bad thing. Probably when they started glueing a bunch of badly executed security crud onto the main product.

      • lolinder 4 hours ago

        The earliest warning sign I had for GitLab was when they eliminated any pricing tier below their equivalent of GitHub's Enterprise tier.

        That day, they very effectively communicated that they had decided they were only interested in serving Enterprises, and everything about their product has predictably degraded ever since, to the point where now they're now branding themselves "the most comprehensive AI-powered DevSecOps Platform" with a straight face.

      • Espressosaurus 6 hours ago

        GitLab can't even show you more than a few lines of context without requiring you to manually click a bunch of times. Forget the CI functionality, for pull requests it's absolutely awful.

      • plagiarist 6 hours ago

        I decided it was a bad thing when they sent password reset emails to addresses given by unauthenticated users. Not that I ever used them. But now it is a hard no, permanently.

        They have since had other also severe CVEs. That has made me feel pretty confident in my decision.

    • danillonunes 12 hours ago

      But you can do the same with GitHub, right? Although most docs and articles focus on 3rd party actions, nothing stops you to just run everything in your own shell script.

      • lolinder 7 hours ago

        Yes, you can, and we do at my current job. Much of the time it's not even really the harder approach compared to using someone else's action, it's just that the existence of third party actions makes people feel obliged to use them because they wouldn't want to be accused of Not Invented Here Syndrome.

    • arccy 7 hours ago

      if anything, gitlab's ci seems even worse...

yunusefendi52 5 minutes ago

I think devbox.sh would solve some of the issues, especially local development. You can also run devbox in CI

spooneybarger 6 hours ago

A lot of folks in this thread are focusing on the monorepo aspect of things. The "Pull request and required checks" problem exists regardless of monorepo or not.

GitHub Actions allows you to only run checks if certain conditions are met, like "only lint markdown if the PR contains *.md files". The moment you decide to use such rules, you have the "Pull request and required checks" problem. No "monorepo" required.

GitHub required checks at this time allow you to use with external services where GitHub has no idea what might run. For this reason, required checks HAVE to pass. There's no "if it runs" step. A required check on an external service might never run, or it might be delayed. Therefore, if GH doesn't have an affirmation that it passed, you can't merge.

It would be wonderful if for jobs that run on GH where GH can know if the action is supposed to run, if required checks could be "require all these checks if they will be triggered".

I have encountered this problem on every non-trivial project I use with GitHub actions; monorepo or not.

  • saxonww 3 hours ago

    This isn't really the problem, though. This is an easy problem to solve; the real problem is that it costs money to do so.

    Also: I'm not asserting that the below is good, just that it works.

    First, don't make every check a required check. You probably don't need to require that linting of your markdown files passes (maybe you do! it's an example).

    Second, consider not using the `on:<event>:paths`, but instead something like `dorny/paths-filter`. Your workflow now runs every time; a no-op takes substantially less than 1 minute unless you have a gargantuan repo.

    Third, make all of your workflows have a 'success' job that just runs and succeeds. Again, this will take less than 1 minute.

    At this point, a no-op is still likely taking less than 1 minute, so it will bill at 1 minute, which is going to be $.008 if you're paying.

    Fourth, you can use `needs` and `if` now to control when your 'success' job runs. Yes, managing the `if` can be tricky, but it does work.

    We are in the middle of a very large migration into GitHub Actions from a self-hosted GitLab. It was something we chose, but also due to some corporate choices our options were essentially GitHub Actions or a massive rethink of CI for several dozen projects. We have already moved into code generation for some aspects of GitHub Actions code, and that's the fifth and perhaps final frontier for addressing this situation. Figure out how to describe a graph and associated completion requirements for your workflow(s), and write something to translate that into the `if` statements for your 'success' jobs.

p1necone 6 hours ago

There's a workaround for the 'pull request and required check' issue. You create an alternative 'no op' version of each required check workflow that just does nothing and exits with code 0 with the inverse of the trigger for the "real" one.

The required check configuration on github is just based off of job name, so either the trigger condition is true, and the real one has to succeed or the trigger condition is false and the no op one satisfies the PR completion rules instead.

It seems crazy to me that such basic functionality needs such a hacky workaround, but there it is.

nunez 14 hours ago

Posts like this make me miss Travis. Travis CI was incredible, especially for testing CI locally. (I agree with the author that act is a well done hack. I've stopped using it because of how often I'd have something pass in act and fail in GHA.)

> GitHub doesn't care

My take: GitHub only built Actions to compete against GitLab CI, as built-in CI was taking large chunks of market share from them in the enterprise.

  • sureIy 8 hours ago

    To be fair, GitHub also charges for Actions minutes and storage, so it's one of the few pieces that do generate revenue.

ripped_britches 3 hours ago

My man/woman - you gotta try buildkite. It’s a bit more extra setup since you have to interface with another company, more API keys, etc. But when you outgrow GH actions, this is the way. Have used buildkite in my last two jobs (big US tech companies) and it has been the only pleasant part of CI.

ryanisnan 6 hours ago

One really interesting omission to this post is how the architecture of GitHub actions encourages (or at the very least makes deceivingly easy) making bad security decisions.

Common examples are secrets. Organization or repository secrets are very convenient, but they are also massive security holes just waiting for unsuspecting victims to fall into.

Repository environments have the ability to have distinct secrets, but you have to ensure that the right workflows can only access the right environments. It's a real pain to manage at scale.

Being able to `inherit` secrets also is a massive footgun, just waiting to leak credentials to a shared action. Search for and leak `AWS_ACCESS_KEY_ID` anyone?

Cross-repository workflow triggering is also a disaster, and in some circumstances you can abuse the differences in configuration to do things the source repository didn't intend.

Other misc. things about GHA also are cool in theory, but fall down in practice. One example is the wait-timer concept of environments. If you have a multi-job workflow using the same environment, wait-timer applies to EACH JOB in the environment. So if you have a build-and-test workflow with 2 jobs, one for build, and one for test, each step will wait `wait-timer` before it executes. This makes things like multi-environment deployment pipelines impossible to use this feature, unless you refactor your workflows.

Overall, I'd recommend against using GHA and looking elsewhere.

  • mdaniel an hour ago

    > Search for and leak `AWS_ACCESS_KEY_ID` anyone?

    Well that's just someone being a dumbass, since AssumeRoleWithWebIdentity (and its Azure and GCP equivalent) have existed for quite a while. It works flawlessly and if someone does do something stupid like `export HURP_DURP=$AWS_ACCESS_KEY_ID; printenv` in a log, that key is only live for about 15 minutes so the attacker better hurry

    Further, at least in AWS and GCP (I haven't tried such a thing in Azure) on can also guard the cred with "if the organization and repo are not ..." then the AssumeRole 403s to ensure that my-awesome-org/junior-dev-test-repo doesn't up and start doing fun prod stuff in GHA

    I hate GHA probably more than most, but one can footgun themselves in any setup

  • wrboyce 3 hours ago

    While I do agree with you regarding encouraging bad secret management practices, one fairly nice solution I’ve landed on is using terraform to manage such things. I guess you could even take it a step further to have a custom lint step (running on GHA, naturally) that disallows secrets configured in a certain manner and blocks a deploy (again, on GHA) on failure.

    I guess what I’m saying is, it’s GHA all the way down.

  • junto 5 hours ago

    Whilst I do detest much of Azure DevOps, one thing I do like about their pipelines is that we can securely use service connections and key vaults in Azure to secure pipeline tasks that require credentials to be managed securely.

  • maccard 5 hours ago

    What’s your suggestion for not-GHA?

posix86 6 hours ago

One thing that sounds very nice about Github are merge queues: Once your PR is ready, rather than merging, you submit it to the merge queue, which will rebase it on the last PR also on the merge queue. It then runs the CI on each PR, and finally merges them automatically once successful. If CI fails, doesn't get merged, and the next PR skips yours on the chain.

Still a lot of computation & some wait time, but you can just click & forget. You can also parallelize it; since branches are rebased on each other, you can run CI in advance and, assuming your predecessor is also successful, reuse the result from yours.

Only available for enterprise orgs though.

  • tialaramex 5 hours ago

    That sounds roughly like what happens for Rust. I write a Rust PR, somebody reviews it, they have feedback, I modify the PR, they're happy, and it passes to bors (originally: https://github.com/graydon/bors) which then tries to see whether this can be merged with Rust and if so does so.

    It is nice to know that if humans thought your change is OK, you're done. I've only committed small changes (compiler diagnostics, documentation) nothing huge, so perhaps if you really get up in it that's more work, but it was definitely a pleasant experience.

    ... and sure enough it turns out that work on one of the bors successors was in fact discontinued because you should just use this GitHub feature. TIL.

bramblerose 21 hours ago

In the end, this is the age old "I built by thing on top of a 3rd party platform, it doesn't quite match my use case (anymore) and now I'm stuck".

Would GitLab have been better? Maybe. But chances are that there is another edge case that is not handled well there. You're in a PaaS world, don't expect the platform to adjust to your workflow; adjust your workflow to the platform.

You could of course choose to "step down" (PaaS to IaaS) by just having a "ci" script in your repo that is called by GA/other CI tooling. That gives you immense flexibility but also you lose specific features (e.g. pipeline display).

  • thayne 5 hours ago

    The problem is that your "ci" script often needs some information from the host system, like what is the target git commit? Is this triggered by a pull request, or a push to a branch? Is it triggered by a release? And if so, what is the version of the release?

    IME, much of the complexity in using Github Actions (or Gitlab CI, or Travis) is around communicating that information to scripts or build tools.

    That and running different tasks in parallel, and making sure everything you want passes.

  • perlgeek 20 hours ago

    > Would GitLab have been better?

    My impression of gitlab CI is that it's also not built for monorepos.

    (I'm a casual gitlab CI user).

    • dezgeg 19 hours ago

      I'm not sure if there's a monorepo vs polyrepo difference; just that anything complex is pretty painful in gitlab. YAML "programming" just doesn't scale.

  • Hamuko 19 hours ago

    Doesn't everything in GitLab go into a single pipeline? GitHub at least makes splitting massive CI/CD setups easier by allowing you to write them as separate workflows that are separate files.

    • dijksterhuis 18 hours ago

      > GitHub at least makes splitting massive CI/CD setups easier by allowing you to write them as separate workflows that are separate files.

      this makes me feel like you’re really asking “can i split up my gitlab CICD yaml file or does everything need to be in one file”.

      if that’s the case:

      yes it does eventually all end up in a single pipeline (ignoring child pipelines).

      but you can split everything up and then use the `include` statement to pull it all together in one main pipeline file which makes dealing with massive amounts of yaml much easier.

      https://docs.gitlab.com/ee/ci/yaml/includes.html

      you can also use `include` to pull in a yaml config from another project to add things like SAST on the fly.

      previous workplace i had like 4 CICD template repos and constructed all 30 odd actual build repos from those four templates.

      used `include` to pull in some yaml template jobs, which i made run when by doing something like (it’s been a while, might get this wrong)

          include:
            project: 'cicd/templates'
            file: 'builds.yml'
      
      
          stages:
            - build
      
          job_a:
            stage: build
            extends: .job_a_from_template
            variables:
              IMAGE_NAME: "myimage"
              IMAGE_REPO: "somerepo.org"
      
      
      this doesn’t run anything for `job_b_from_template` … you just end up defining the things you want to run for each case, plus any variables you need to provide / override.

      you can also override stuff like rules on when it should run if you want to. which is handy.

      gitlab CICD can be really modular when you get into it.

      if that wasn’t the case: on me.

      edit: switched to some yaml instead of text which may or may not be wrong. dunno. i have yet to drink coffee.

      • dijksterhuis 15 hours ago

        addendum you can also do something like this, which means you don’t have to redefine every job in your main ci file, just define the ones you don’t want to run

            include:
              project: 'cicd/templates'
              file: 'builds.yml'
        
            variables:
              IMAGE_NAME: something
              IMAGE_REPO: some.org
        
            job_b:
              rules:
                - when: never
        
        where the template you import has a job_a and job_b definition. both get pulled in, but job_b gets overwritten so it never runs.

        less useful when just splitting things into multiple files to make life simpler.

        super useful when using the same templates across multiple independent repositories to make everything build in as close to the same way as possible.

    • dezgeg 19 hours ago

      You can have pipelines trigger child pipelines in gitlab, but usability of them is pretty bad, viewing logs/results of those always needs extra clicking.

zxor 7 hours ago

> The problem is that this step will only run when I change something in the web-app1 folder. So if my pull request only made changes in api1 I will never be able to merge my pull request!

This just seems like a bad implementation to me?

There are definitely ways to set up your actions so that they run all of the unit tests without changes if you'd like, or so that api1's unit tests are not required for a web-app1 related PR to be merged.

  • hightrix 6 hours ago

    Absolutely correct. When creating a new workflow, I always disable push/pull_request triggered builds and instead use the manually triggered `workflow_dispatch` method. This makes testing a new workflow much easier.

    Additionally, you can use conditionals based on inputs in the `workflow_dispatch` meaning that you could easily setup a "skip api tests" or "include web tests" option.

    • verdverm 4 hours ago

      It sounds like they have the logic to skip certain things if nothing has changed. The problem is around pull request gates and the lack of dynamic "these tests must be passing before merging is allowed". There are setting on a repository in the ruleset / status checks area that are configured outside of the dynamic yaml of the GHA workflow

bhaney 18 hours ago

Article title: "[Common thing] doesn't work very well!"

Article body: "So we use a monorepo and-"

Tale as old as time

  • rasso 7 hours ago

    I‘m also struggling with gh actions. And none of my repos is a monorepo.

SamuelAdams 7 hours ago

> Our code sits in a monorepo which is further divided into folders. Every folder is independent of each other and can be tested, built, and deployed separately.

If this is true, and you still have problems running specific Actions, why not break this into separate repositories?

  • verdverm 4 hours ago

    There is a mono vs poly repo tradeoff. Pros & cons to each approach. If you are doing monorepo, it would be antithetical to break it up into the poly paradigm. You really don't want both

  • Always42 4 hours ago

    My immediate response as well.

androa 20 hours ago

GitHub (Actions) is simply not built to support monorepos. Square peg in a round hole and all that. We've opted for using `meta` to simulate monorepos, while being able to use GitHub Actions without too much downsides.

  • justin_oaks 5 hours ago

    Which makes me wonder if there is a way to simulate multiple repos while maintaining a mono repo. Or mirror a portion of a monorepo as a single repo.

    Obviously this would be a real pain to implement just to fix the underlying problem, but it's an interesting (awful) solution

  • Imustaskforhelp 19 hours ago

    hey could you please share the`meta` tool you mentioned , sounds interesting ! couldn't find it on internet [skill issue]

    • joshka 19 hours ago

      Guessing it's https://github.com/mateodelnorte/meta googlefu "meta github repo"

      • Imustaskforhelp 19 hours ago

        hey thanks!

        definitely interesting!

        I do wonder if this really solves the author problem because by the looks of it , you just have to run meta command and it would run over each of the sub directory. While at the same time , I think I like it because this is what I think people refer to as "modular monolith"

        Combining this with nats https://nats.io/ (hey if you don't want it to be over the network , you could use nats with the memory model of your application itself to reduce any overhead) and essentially just get yourself a really modular monolith in which you can then seperate things selectively (ahem , microservices) afterwards rather easily.

        • HumanOstrich 18 hours ago

          Modular monolith refers to the architecture of your application[1]. It's a different concept from "monorepo", although they can be used together.

          I'm not sure what NATS has to do with anything in this post or discussion. Also, a modular monolith is almost the antithesis of microservices.

          [1]: https://www.thoughtworks.com/en-us/insights/blog/microservic...

xinayder 12 hours ago

I tried to use GitHub Actions on Forgejo and... It's so much worse than using an actual CI pipeline.

With Woodpecker/Jenkins you know exactly what your pipeline is doing. With GitHub actions, not even the developers of the actions themselves know what the runner does.

  • lolinder 7 hours ago

    > use GitHub Actions on Forgejo

    What does this even mean? Are you talking about Forgejo Actions, or are you somehow hosting your code on a Forgejo instance but running CI through GitHub?

    > With Woodpecker/Jenkins you know exactly what your pipeline is doing.

    If you wrote it from the ground up, sure. On the other hand, I've inherited Jenkins pipelines that were written years before I got there and involved three to four different plugins, and they're way worse to work with than the GitHub Actions that I inherited.

  • myaccountonhn 7 hours ago

    Best one I've used is the CI of sourcehut. So simple and so damn easy to set up.

    • gabeio 6 hours ago

      You basically achieve the same result on github actions if you just ignore all of the github action yaml “magic” settings in the syntax and let your makefile/script do the logic which also makes it trivial to debug locally. But upvote because I do love sourcehut, it’s just so clean!

tevon 21 hours ago

I call writing GitHub Actions "Search and Deploy", constantly pushing to a branch to get an action to run is a terrible pattern...

You'd think, especially with the deep VS Code integration, they'd have at least a basic sanity-check locally, even if not running the full pipeline.

  • 8n4vidtmkvmk 19 hours ago

    Not just me then? I was trying to fix a GitHub action just today but I have no clue how I'm supposed to tear it, so I just keep making tiny changes and pushing.... Not a good system but I'm still within the free tier so I'm willing to put up with it I guess.

    • arccy 7 hours ago

          git commit --allow-empty -m "bump ci"
      
      unless your pipeline does magic with trying to detect changed files
      • mdaniel an hour ago

        If this is to troubleshoot non-code related failures (perm issues, connection timed out, whatever influences success that doesn't require a code change) then surely the repo's history would benefit from one just clicking "Re-run Job", or its equivalent $(gh ...) invocation, right?

    • masklinn 18 hours ago

      I think it’s everyone, debugging GH actions is absolute hell, and it gets terrifying when the action interacts with the world (e.g. creating and deploying packages to a registry).

      • oefrha 18 hours ago

        > it gets terrifying when the action interacts with the world (e.g. creating and deploying packages to a registry).

        To be fair, testing actions with side effects on the wider world is terrifying even if you’re running it locally, maybe more so because your nonstandard local environment may have surprises (e.g. an env var you set then forgot) while the remote environment mostly only has stuff you set/installed explicitly, and you can be sloppier (e.g. accidentally running ./deploy when you wanted to run ./test). That part isn’t a GH Actions problem.

  • pavon 7 hours ago

    Ah yes, I have a git alias created specifically for the "we don't know what it does until we push it" world of CI:

    > yolo = "!git commit --all --amend --no-edit && git push --force #"

  • nunez 14 hours ago

    Biggest pet peeve of GHA by a country mile.

verdverm 4 hours ago

Monorepos come with a different set of tradeoffs from polyrepos. Both have their pains. We have a similar setup with Jenkins, and have used CUE to tame a number of these issues. We did so by creating (1) a software catalog (2) per-branch config for versions and CI switches

Similarly, we are adopting Dagger, more as part of a larger "containerize all of our CI steps" which works great for bringing parity to CI & local dev work. There are a number of secondary benefits and the TUI / WUI logs are awesome.

Between the two, I have removed much of the yaml engineering in my work

OptionOfT 12 hours ago

So the way I've solved the multiple folders with independent checks is like this:

    all-done:
      name: All done
      # this is the job that should be marked as required on GitHub. It's the only one that'll reliably trigger
      # when any upstream fails: success
      # when all upstream skips: pass
      # when all upstream success: success
      # combination of upstream skip and success: success
      runs-on: ubuntu-latest
      needs:
        - calculate-version
        - cargo-build
        - cargo-fmt
        - cargo-clippy-and-report
        - cargo-test-and-report
        - docker-build
        - docker-publish
      if: |
        always()
      steps:
        - name: Fail!
          shell: bash
          if: |
            contains(needs.*.result, 'failure') ||
            contains(needs.*.result, 'cancelled')
          run: |
            echo "One / more upstream failed or was cancelled. Failing job..."
  
            exit 1
  
        - name: Success!
          shell: bash
          run: |
            echo "Great success!"

That way it is resilient against checks not running because they're not needed, but it still fails when any upstream actually fails.

Now, I did end up running the tests of the front-end and back-end because they upload coverage, and if my coverage tool doesn't get both, it'll consider it as a drop in coverage and fail its check.

But in general, I agree with the writer of the post that it all feels like it's not getting enough love.

For example, there is no support for yaml anchors, which really hampers reusability on things that cannot be extracted to separate flows (not to mention separate flows can only be nested 4 deep).

There is also the issue that any commit made by GitHub actions doesn't trigger another build. This is understandable, as you want to avoid endless builds, but sometimes it's needed, and then you need to do the ugly workaround with a PAT (and I believe it can't even be a fine-grained one). Combine that with policies that set a maximum time limit on tokens, your build becomes brittle, as now you need to chase down the person with admin access.

Then there is the issue of Docker actions. They tell you to pin the action to an sha to prevent replacements. Except the action itself points to a replaceable tag.

Lastly, there is a bug where when you create a report for your action, you cannot specify the parent it belongs to. So your ESLint report could be made a child of your coverage report.

hinkley 6 hours ago

Re: monorepo

> In GitHub you can specify a "required check", the name of the step in your pipeline that always has to be green before a pull request is merged. As an example, I can say that web-app1 - Unit tests are required to pass. The problem is that this step will only run when I change something in the web-app1 folder. So if my pull request only made changes in api1 I will never be able to merge my pull request!

Continuous Integration is not continuous integration if we don’t test that a change has no deleterious side effects on the rest of the system. That’s what integration is. So if you aren’t running all of the tests because they’re slow, then you’re engaging in false economy. Make your tests run faster. Modern hardware with reasonable test runners should be able to whack out 10k unit tests in under a minute. The time to run the tests goes up by a factor of ~7-10 depending on framework as you climb each step in the testing pyramid. And while it takes more tests to cover the same ground, with a little care you can still almost halve the run time replacing one test with a handful of tests that check the same requirement one layer down, or about 70% moving down two layers.

One thing that’s been missing from most of the recent CI pipelines I’ve used is being able to see that a build is going to fail before the tests finish. The earlier the reporting of the failure the better the ergonomics for the person who triggered the build. That’s why the testing pyramid even exists.

  • rustd 5 hours ago

    Agreed, most of the CI tools don't help in getting feedback early to the developers. I shouldn't have to wait hours for my CI job to complete. Harness is a tool that can reduce build times by caching build artifacts, docker layers and only running a subset of tests that were impacted the by the code change.

  • maccard 5 hours ago

    I agree hardware should be that quick, but CI and cloud hardware is woefully underpowered unless you actively seek it out. I’ve also never seen a test framework spew out even close to that in practice. I’m not even sure most frameworks would do that with noop tests, which is sad.

    • hinkley 4 hours ago

      10 years ago my very testing-competent coworker had us running 4200 tests in 37 seconds. In NodeJS. We should be doing as well that today without a gifted maintainer.

zzo38computer 6 hours ago

I do not use GitHub Actions for these purposes, and if I did, I would want to ensure that it is a file that can run locally or whatever else just as well. I don't use GitHub Actions to prevent pull requests from being merged (I will always manage them manually), and do not use GitHub Actions to manage writing the program, for testing the program (it would be possible to do this, but I would insist on doing it in a way that is not vendor-locked to GitHub, and by putting most of the stuff outside of the GitHub Actions file itself), etc.

I do have a GitHub Actions file for a purpose which is not related to the program itself; specifically, for auto-assignment of issues. In this case, it is clearly not intended to run locally (although in this case you could do so anyways if you could install the "gh" program on your computer and run the command mentioned there locally, but it is not necessary since GitHub will do it automatically on their computer).

  on:
    issues:
      types:
        - opened
    pull_request:
      types:
        - opened
  permissions:
    contents: read
    issues: write
    pull-requests: write
  jobs:
    default:
      runs-on: ubuntu-latest
      steps:
        - run: gh issue edit ${{ github.event.issue.number }} --add-assignee ${{ github.repository_owner }}
          env:
            GH_TOKEN: ${{ github.token }}
            GH_REPO: ${{ github.repository }}
flohofwoe 19 hours ago

IMHO the main problem with GH Actions is that the runners are so slow. Feels like running your build on a frigging C64 sometimes ;)

  • rustd 5 hours ago

    GH hosted runners use shared hardware so the performance is never good. There are quite a few options available. Harness CI offers hyper optimized build infrastructure, paired with software intelligence (caching, running subset of tests based on the code change) can reduce build times up-to 4X compared to GH Actions.

  • ramon156 18 hours ago

    Blacksmith is your buddy. Its free and just has better images for single-core operations. Unless you're Google, I can guarantee it's faster.

  • Hamuko 19 hours ago

    Are you hosting your own runners or relying on GitHub's?

forty 7 hours ago

For the first point, some mono repo orchestrators (I'm thinking of at least pnpm) have a way to do : run all the (for example) tests for all the packages that had change from master branch + all packages that depend transitively from those packages.

It's very convenient and avoid having to mess with the CI limitations on the matter

rednafi 18 hours ago

You can't run AWS lambda or DyanmoDB locally too (well you can but it's a hassle). So by that logic, we shouldn't use them at all. I don't like working with CI too but I'll take GitHub Actions over Jenkins/CircleCI/TravisCI any day.

  • chriswarbo 17 hours ago

    > You can't run AWS lambda or DyanmoDB locally too (well you can but it's a hassle). So by that logic, we shouldn't use them at all.

    No, applying the logic to something like Lambda would mean implementing handlers like:

        function handle(lambdaRequest, lambdaContext) {
          return myFunction(
            stuffExtractedFromLambdaRequest(lambdaRequest),
            stuffExtractedFromLambdaContext(lambdaContext)
          );
        }
    
    Then there's no need to go through the hassle of running Lambda functions locally; since we can just run `myFunction` locally instead.

    Dynamo isn't the same, since it's just a service/API that we call; we don't implement its logic, like we do for CI tasks, Lambda functions, etc.

    Whilst you're right that it's a hassle to run DynamoDB locally (although not too bad, in my experience); that's also not necessary. It's fine to run code locally which talks to a remote DynamoDB; just set up the credentials appropriately.

    • rednafi 17 hours ago

      Yeah, and that doesn't stop us from using either of them. What I tried to convey is that GHA isn't ideal and it has a few warts but it's still better than most of the options available out there.

  • sunshowers 8 hours ago

    The problem with the analogy is that GHA's interface is quite thick.

dalton_zk 4 hours ago

One options its create your own CI, I think the others tools have pros/cons.

This month I start to create to my team our own tool to build CI, I'm using go lang and create a webhook who call my API and apply what is need.

I'm saying this because you can create the CI with your features.

kjuulh 19 hours ago

I am biased because I built the rust SDK for dagger. But I think it is a real step forward for CI. Is it perfect? Nope. But it allows fixing a lot of the shortcomings the author has.

Pros:

- pipeline as code, write it as golang, python, typescript or a mix of thr above.

- Really fast once cached

- Use your languages library for code sharing, versioning and testing

- Runs everywhere local, ci etc. Easy to change from github actions to something else.

Cons:

- Slow on the first run. Lots of pulling of docker images

- The DSL and modules can feel foreign initially.

- Modules are definitely a framework, I prefer just building having a binary I can ship (which is why the rust SDK doesnt support modules yet).

- Doesn't handle large mono repos well, it relies heavily on caching and currently runs on a single node. It can work if you don't have 100 of services especially if the builder is a large machine.

Just the fact that you can actually write ci pipelines that can be tested, packaged, versioned etc. Allows us to ship our pipelines as products which is quite nice and something we've come to rely on heavily

  • electromech 6 hours ago

    I'm genuinely intrigued by Dagger, but also super confused. For example, this feels like extra complexity around a simple shell command, and I'm trying to grok why the complexity is worth it: https://docs.dagger.io/quickstart/test/#inspect-the-dagger-f...

    I'm a fanboy of Rust, Containerization, and everything-as-code, so on paper Dagger and your Rust SDK seems like it's made for me. But when I read the examples... I dunno, I just don't get it.

    • mdaniel 5 hours ago

      I'm merely an outsider to Dagger, but I believe the page you linked to would give one the impression "but why golang[1] around some shell literals?!" because to grok its value one must understand that m.BuildEnv(source) <https://docs.dagger.io/quickstart/env#inspect-the-dagger-fun...> is programmatically doing what https://docs.github.com/en/actions/writing-workflows/workflo... would do: define the docker image (if any), the env vars (if any), and other common step parameters

      That conceptually allows one to have two different "libraries" in your CI: one in literal golang, as a function which takes in a source and sets up common step configurations, and the other as Dagger Functions via Modules (<https://docs.dagger.io/features/modules> or <https://docs.dagger.io/api/custom-functions#initialize-a-dag...>) which work much closer to GHA uses: blocks from a organization/repo@tag style setup with the grave difference that they can run locally or in CI, which for damn sure is not true of GHA uses: blocks

      The closest analogy I have is from GitLab CI (since AFAIK GHA does not allow yaml anchors nor "logical extension"):

        .common: &common # <-- use whichever style your team prefers
          image: node:21-slim
          cache: {} # ...
          environment: [] # ...
        my-job1:
          stage: test
          <<: *common
          script: [ npm, run, test:unit, run ]
        my-job2:
          stage: something-else
          extends: .common
          script: echo "do something else"
      
      1: I'm aware that I am using golang multiple times in this comment, and for sure am a static typing fanboi but as the docs show they allow other languages, too
kazinator 4 hours ago

> My team consists of about 15 engineers

If it's not open source, I have no idea why you'd use GitHub at all. (And even then.)

Keep your eggs in your own nest.

joshdavham 19 hours ago

> It's a known thing that there is no way of running GitHub Actions locally. There is a tool called act but in my experience it's subpar.

I really hope there will be a nice, official tool to run gh actions locally in the future. That would be incredible.

spzb 7 hours ago

My recent experience with Github Actions is that it will randomly fail running a pipeline that hasn't changed with an incomprehensible error message. I re-run the action a few hours later and it works perfectly.

  • quesera 4 hours ago

    This is great.

    I also enjoy the "randomly and undebuggably hang until timeout" (25mins?) which is annoying and incomprehensible and costs money.

    Rerunning the same jobs always passes.

rickette 15 hours ago

Every CI system has its flaws but GitHub Actions in my opinion is pretty nice especially in terms of productivity; easy to setup, tons of prebuild actions, lots of examples, etc.

I've used Tekton, Jenkins, Travis, Hudson, StarTeam, Rational Jazz, Continuum and a host of other CI systems over the years but GitHub Actions ain't bad.

esafak 4 hours ago

Where did the terrible idea of pipelines as config come from anyway?

keybored 8 hours ago

Why is this so difficult?

1. We apparently don’t even have a name for it. We just call it “CI” because that’s the adjacent practice. “Oh no the CI failed”

2. It’s conceptually a program that reports failure if whatever it is running fails and... that’s it

3. The long-standing principle of running “the CI” after merging is so backwards that that-other Hoare disparagingly called the correct way (guard “main” with a bot) for The Not Rocket Science Principle or something. And that smug blog title is still used to this day (or “what bors does”)

4. It’s supposed to be configured declaratively but in the most gross way that “declarative” has ever seen

5. In the true spirit of centralization “value add”: the local option of (2) (report failure if failed) has to be hard or at the very least inconvenient to set up

I’m not outraged when someone doesn’t “run CI”.

makingstuffs 20 hours ago

Not sure if I am missing something but you can definitely run (some?) GH actions locally with act: https://github.com/nektos/act

Seen a couple posts on here say otherwise.

  • alhadrad 19 hours ago

    Act has limitations because GitHub Actions run via virtualization, while Act runs via containerization. This means that actions behave differently across the two platforms.

  • joshdavham 19 hours ago

    He mentioned act in the article.

bitliner2 18 hours ago

Welcome to the jungle.

https://medium.com/@bitliner/why-gitlab-can-be-a-pain-ae1aa6...

I think it’s not only GitHub.

Ideally we should handle it as any other code, that is: do tests, handle multiple environments including the local environment, lint/build time error detection etc

  • rustd 5 hours ago

    Agree. CI should be handled as code.

  • benrutter 17 hours ago

    Unhappy to confirm that for any poor souls using Azure DevOps, it's even worse.

vrnvu 20 hours ago

> GitHub doesn't care

GitHub cares. GitHub cares about active users on their platform. Whether it's managing PRs, doing code reviews, or checking the logs of another failed action.

  • joshka 19 hours ago

    GitHub often actively doesn't act in situations where acting would be prudent, which portrays from an outside perspective a disinterest in those who give their time to document shortcomings. Would you care to guess when the last time that the GitHub API was updated? It's probably much longer than you'd think (2+ years at this point).

  • presentation 18 hours ago

    They don’t care about things that I care about, including everything the author talked about, and also things like allowing whitespace-ignore on diffs to be set on by default in a repo or per user - an issue that’s been open for half a decade now.

    (Whitespace is just noise in a typescript repo with automatic formatting)

    https://github.com/orgs/community/discussions/5486

angoragoats 15 hours ago

Why is this team sticking multiple directories that are “independent of each other” into a single repository? This sounds like a clear case of doing version control wrong. Monorepos come with their own set of challenges, and I don’t think there are many situations where they’re actually warranted. They certainly don’t help for completely independent projects.

  • ripped_britches 3 hours ago

    No project within a single organization is completely independent. In general they all serve to meet a unified business objective and developers often need the global context on occasion.

    I used to be a multirepo proponent, but have fallen in love with Bazel and “everything bagel” repo.

  • cbare 7 hours ago

    Yeah, sounds like the problems are more due to monorepos rather than with GitHub actions. Seems like the pendulum always swings too far. Overdoing microservices results in redundant code and interservice spaghetti. Monorepos have their own set of issues. The only solution is to think carefully about a what size chunk of functionality you want to build, test, and deploy as a unit.

pshirshov 16 hours ago

> Jenkins, TeamCity

Yeah-yeah, but it's not like they allow you to run your build definitions locally nor they address some other concerns. With GHA you may use nix-quick-install in a declarative manner, nixify your builds and then easily run them locally and under GHA. In case of jenkins/tc you would have to jump through much more hoops.

aa-jv 17 hours ago

I use Github Actions as a fertile testing playground to work out how to do things locally.

For example, if you've ever had to wade into the codesigning/notarization quagmire, observing the methods projects use with Github Actions to do it, can teach you a lot about how to do things, locally.

alkonaut 18 hours ago

That GH Actions and Azure Pipelines both settled for this cursed Yaml is hard to understand. Just make a real programming language do it! And ffs make a local test env so I can run the thing.

rustd 5 hours ago

[flagged]

  • Celeo 5 hours ago

    Op is lead for this product.