Skip to content

coveo/devops-coding-challenge

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 

Repository files navigation

Coveo DevOps Challenge

The Challenge

Your challenge, should you choose to accept it, is to develop an AWS S3 storage analysis tool. To test your tool, you will have to create a free Amazon account (if you don't already have one).

Specifications

The tool is a shell command line utility that returns information over all S3 buckets in an Amazon account.

  • Your tool must work on Linux, OSX and Windows.
  • It must be easy to install and use.
  • Ideally, your tool won't require installation of any other tools / libraries / frameworks to work.
  • Time is money, we cannot afford running a tool that takes hours to complete, your solution should return results within seconds (or minutes if you are willing to test our patience :-).

Time allowed

People who have successfully passed the challenge and are now happy members of the Coveo team usually took from 4 to 10 hours to complete it.

Remember the KISS principle...

The tool must return the following information

For each bucket:

  • Name
  • Creation date
  • Number of files
  • Total size of files
  • Last modified date of the most recent file
  • And the most important of all, how much does it cost

The following options should be supported

  • Display

    • Ability to get the size results in bytes, kB, MB, ...
    • Ability to group buckets by regions
  • Filters

    • By bucket name
    • By storage type (Standard, IA, RR). You may provide stats on the objects in the buckets (how many have which storage type) and/or add a filter on storage type (the bucket data would only reflect the objects that have the chosen storage type).

Some additional features that could be useful (optional)

It would be nice to:

  • Support prefixes, glob and/or regexes in the bucket filter (e.g.: s3://mybucket/Folder/SubFolder/log*).
  • Organize the results according to the encryption type
  • Get additional buckets information (life cycle, cross-region replication, etc.)
  • Take into account the previous file versions in the count + size calculation.

Some statistics to check the percentage of space used by a bucket, or any other good ideas you could have, are more than welcome.

Rules

  • Your are free to use the programming language and the SDK of your choice, but remember that installation must not require us to install external tools in order to test the result of your work.
  • We will test your work in our environment (which contains several millions of files). The overall performance of your tool will be evaluated. Most of the projects we receive take weeks to run in our environment, can you do better?
  • Your code must be made available on GitHub or any other public version control software. Your project must be a standalone project (i.e. do not fork it from our challenge or any other project).

Advice

  • Try to design and implement your solution as you would do for real production code. Show us how you create clean, maintainable code that does awesome stuff. Build something that we'd be happy to contribute to. This is not a programming contest where dirty hacks win the game.
  • Feel free to add more features! Really, we're curious about what you can think of. We'd expect the same if you worked with us.
  • Documentation and maintainability is a plus.
  • Don't you forget those unit tests.
  • We don’t want to know if you can do exactly as asked (or everybody would have the same result). We want to know what you bring to the table when working on a project, what is your secret sauce. More features? Best solution? Thinking outside the box?

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published