HTML5 Boilerplate homepage | Documentation
table of contents

Miscellaneous

.gitignore

HTML5 Boilerplate includes a basic project-level .gitignore. This should
primarily be used to avoid certain project-level files and directories from
being kept under source control. Different development-environments will
benefit from different collections of ignores.

OS-specific and editor-specific files should be ignored using a “global
ignore” that applies to all repositories on your system.

For example, add the following to your ~/.gitconfig, where the .gitignore
in your HOME directory contains the files and directories you'd like to
globally ignore:

[core]
    excludesfile = ~/.gitignore

.editorconfig

The .editorconfig file is provided in order to encourage and help you and
your team define and maintain consistent coding styles between different
editors and IDEs.

By default, .editorconfig includes some basic
properties that reflect the
coding styles from the files provided by default, but you can easily change
them to better suit your needs.

In order for your editor/IDE to apply the
properties from the
.editorconfig file, you will need to install a
plugin
.

N.B. If you aren't using the server configurations provided by HTML5
Boilerplate, we highly encourage you to configure your server to block
access to .editorconfig files, as they can disclose sensitive information!

For more details, please refer to the EditorConfig
project
.

Server Configuration

H5BP includes a .htaccess file for the Apache HTTP server. If you are not using
Apache as your web server, then you are encouraged to download a
server configuration that corresponds
to your web server and environment.

Servers and Stacks

A comprehensive list of web servers and stacks are beyond the scope of this
documentation, but some common ones include:

.htaccess

A .htaccess (hypertext access) file is a
Apache HTTP server configuration file.
The .htaccess file is mostly used for:

If you have access to the main server configuration file (usually called
httpd.conf), you should add the logic from the .htaccess file in, for
example, a section in the main configuration file. This is usually
the recommended way, as using .htaccess files slows down Apache!

To enable Apache modules locally, please see:
https://github.com/h5bp/server-configs-apache/wiki/How-to-enable-Apache-modules.

In the repo the .htaccess is used for:

When using .htaccess we recommend reading all inline comments (the rules after
a #) in the file once. There is a bunch of optional stuff in it.

If you want to know more about the .htaccess file check out
https://httpd.apache.org/docs/current/howto/htaccess.html.

Notice that the original repo for the .htaccess file is this
one
.

crossdomain.xml

The cross-domain policy file is an XML document that gives a web client —
such as Adobe Flash Player, Adobe Reader, etc. — permission to handle data
across multiple domains, by:

e.g. If a client hosts content from a particular source domain and that
content makes requests directed towards a domain other than its own, the remote
domain would need to host a cross-domain policy file in order to grant access
to the source domain and allow the client to continue with the transaction.

For more in-depth information, please see Adobe's cross-domain policy file
specification
.

robots.txt

The robots.txt file is used to give instructions to web robots on what can
be crawled from the website.

By default, the file provided by this project includes the next two lines:

If you want to disallow certain pages you will need to specify the path in a
Disallow directive (e.g.: Disallow: /path) or, if you want to disallow
crawling of all content, use Disallow: /.

The /robots.txt file is not intended for access control, so don't try to
use it as such. Think of it as a “No Entry” sign, rather than a locked door.
URLs disallowed by the robots.txt file might still be indexed without being
crawled, and the content from within the robots.txt file can be viewed by
anyone, potentially disclosing the location of your private content! So, if
you want to block access to private content, use proper authentication instead.

For more information about robots.txt, please see: