Can I use the packit service as soon as I install it into my repository?

Thanks for your interest in Packit Service! In order to start using the service you need to be whitelisted, which is an action to be done by us. Once we put you on the whitelist, we’ll get in touch with you. We are now on-boarding Fedora contributors (with a Fedora Account System account).

Can I use packit service for any GitHub repository?

Since packit service builds your PRs in Fedora COPR build service, by using Packit-as-a-service, your software needs to comply with COPR rules. If any of these points are violated, we’ll remove the builds and may put you on a blacklist so you won’t be able to use the service again.

How can I contact you?

If you encounter a problem while using Packit-as-a-service, please open an upstream issue. In case of any other questions, feel free to contact us:

Why do I have to maintain .packit.yaml and a spec file upstream?

We are working on simplifying the .packit.yaml so it’s as small as possible. We will also handle all potentially backward incompatible changes of .packit.yaml. Spec file can be downloaded (see specific question below) from Fedora Pagure instead of having it included in the upstream repository.

But what are the benefits?

Packit makes it trivial to run your project as part of an OS. It provides feedback to your project at the time when the changes are being developed so you can fix incompatible code when you are working on it, not when it’s already released. When you push commits to a pull request, you’ll get RPM build and test results right away.

Why Fedora?

We’ve started with Fedora Linux because we work for Red Hat and we ❤ Fedora.

How is Packit different from other services?

Packit connects the existing services (Copr, Pagure, Koji, Bodhi) together.

Can we use Packit with Gitlab?

Packit service is closely tied to GitHub right now because most of the projects are hosted on GitHub. Let us know if you’d like to see Gitlab support.

How can I download RPM spec file if it is not part of upstream repository?

If you do not want to have the RPM spec file in your upstream repository, you can download it in actions section.

Add actions section to your packit.yaml configuration file and download the spec file in a hook post_upstream_clone. Packit service has a limited set of commands available so please use wget or curl.

The configuration file with downloading the RPM spec file now looks like this:

specfile_path: packit.spec
  - packit.spec
  - .packit.yaml
upstream_package_name: packitos
downstream_package_name: packit
    post-upstream-clone: "wget https://src.fedoraproject.org/rpms/packit/raw/master/f/packit.spec"

I have a template of a spec file in my repo: can packit work with it?


The solution is, again, actions and hooks. Just render the spec after the upstream repo is cloned:

specfile_path: my-project.spec
upstream_package_name: my-project-src
downstream_package_name: my-project
    post-upstream-clone: "make generate-spec"

Where the “generate-spec” make target could look like this:

    sed -e 's/@@VERSION@@/$(VERSION)/g' my-project.spec.template >my-project.spec

As a practical example, cockpit-podman project is using this functionality.

Can I use CentOS Stream with packit service?

Yes, you can! It’s very simple, just add centos-stream-x86_64 as a target for the copr_build job:

- job: copr_build
  trigger: pull_request
      - centos-stream-x86_64

After adding tests I see error ‘No FMF metadata found.’

If you encounter this error when running tests via Testing Farm, it means you forgot to initialize the metadata tree with fmf init and include the .fmf directory in the pull request. See Testing Farm documentation for more information.

Does packit work with rpmautospec?

Good that you ask. It does, packit works with rpmautospec quite nicely.

Before you start, please make sure that you follow latest documentation for rpmautospec.

rpmautospec utilizes two RPM macros:

  1. autorel — to populate Release
  2. autochangelog — to figure out changelog

If you want your upstream spec file to also work well when rpmautospec-rpm-macros is not installed, set Release to this:

Release:  %{?autorel}%{!?autorel:1}

This construct uses autorel macro if it’s defined, and if it’s not, it sets release to 1.

For %changelog, you don’t need to include the changelog file upstream and you can have it downstream only, which makes sense - changelog is specific to a release.

How do I install dependencies for my commands in packit-service?

We are running all commands, defined by you, in a sandbox which is locked-down. At the moment we don’t have any mechanism for you to define the dependencies you need and us making them available for you.

In the mean time we are solving these requests one by one, so please reach out to us.

A command failed in packit-service: how do I reproduce it locally?

We don’t have an end-to-end solution to this, yet.

In the meantime, you can pull our production sandbox image and run the command inside. As an example, this is how we were debugging build problems with anaconda:

  1. Clone the upstream git repo.

  2. Pull the sandbox image:

$ docker pull docker.io/usercont/sandcastle:prod
  1. Launch the container and bind-mount the upstream project inside:
$ docker run -ti --rm -v $PWD:/anaconda docker.io/usercont/sandcastle:prod bash
  1. Run commands of your choice:
[root@4af5dbd9c828 /]# cd /anaconda

[root@4af5dbd9c828 anaconda]# ./configure
checking for a BSD-compatible install... /usr/bin/install -c
checking whether build environment is sane... yes
checking for a thread-safe mkdir -p... /usr/bin/mkdir -p
checking for gawk... gawk
checking whether make sets $(MAKE)... yes
checking whether make supports nested variables... yes
checking whether UID '0' is supported by ustar format... yes
checking whether GID '0' is supported by ustar format... yes
checking how to create a ustar tar archive... gnutar
checking whether make supports nested variables... (cached) yes
checking whether make supports the include directive... yes (GNU style)
checking for gcc... gcc
checking whether the C compiler works... yes

Our deployment is running in OpenShift Online which is using docker as a container engine, that’s why we are using docker here and not podman.