Skip to content

Latest commit

 

History

History
131 lines (67 loc) · 13.6 KB

README.md

File metadata and controls

131 lines (67 loc) · 13.6 KB

Global is the New Local: FPGA Architecture at 5nm and Beyond

DOI

This repository holds the source code used to produce the results of the paper entitled "Global is the New Local: FPGA Architecture at 5nm and Beyond".

Prerequisites

In order to reproduce the main results, Synopsys HSPICE (version 2013.12 was used in the paper) and VTR 8.0 are needed. The corresponding commands are specified in setenv.py. The same file contains two variables specifying the number of parallel threads to be used for SPICE simulations, and for the remaining experiments (mainly VPR).

For running the Python scripts, Python 2.7 is required, along with the networkx package. Matplotlib is used for graph plotting.

Additional

Because the routing-resource graphs can consume a lot of space, the scripts compress them using lz4 by default.

Code Organization and Result Reproduction

All scripts should be run from the directory of their source file.

The source code is structured as follows:

SPICE Models

Directory spice_models contains transistor models for the 16nm node (F16), taken from PTM, as well as instructions on how to update the appropriate ASAP7 models to obtain the models used in the paper for the scaled technologies (F7--F3). Before running the experiments, the template files in this directory should be updated according to the instructions.

Wire and Via Parameters

Directory rc_calculations contains the scripts modeling the per-micrometer-length resistance and capacitance of wires as well as resistance of vias. It also contains the scripts to generate the results and produce the tables appearing in the submitted version of the paper, that can be run without any arguments (generate_table_2.py, generate_table_3.py, generate_table_4.py).

The produced results should be included in tech.py (Those produced at the time of writing the paper are already embedded in the file).

Logic Delays

Scripts for delay scaling between technology nodes, fanout-of-4-inverter (FO4) delay simulation, and generating the appropriate tables as they appeared in the submitted version of the paper (generate_table6.py and generate_table7.py are contained in logic_delays/.

The obtained scaled LUT and FF delays should be included in tech.py (Those produced at the time of writing the paper are already embedded in the file).

Wire Delay Measurement

The script for generating the SPICE netlists and measuring the delay of feedback intracluster wires (from the output of one LUT to inputs of other LUTs in the same cluster) and the delay of distributing an intercluster signal from the global wires to the LUT inputs is contained in wire_delays/local_wires.py. It provides various options, like repeater insertion and layer planning, described in the docstrings of the file.

The same directory also contains a script that models global wire delay in a simplified manner, so as to appropriately size the drivers and determine the maximum logical lengths of wires in a given technology and with a given cluster size (global_wires.py]).

Besides the delay modeling scripts, the directory also contains the scripts to generate the required numbers and plot the figures representing feedback delays in various settings, as used in the paper. To generate the data, run get_figure_data.py, without any arguments, which will produce additional Python files, containing the data, tagged with appropriate labels for plotting.

To plot the figures, run python plot_graphs_11_13.py --data figure_#, where # is the number of the previously produced Python file (no extension is required).

Buffer Sizing

As stated in the previous section, buffer sizing is performed by the delay modeling scripts in wire_delays/. The sizes, as well as the measured feedback delays are later used for complete architecture generation, stored in wire_delays/buf_cache/. To store the sizes for local wire drivers, run build_local_buf_cache.py, without arguments. Sizes of global wire drivers will be stored upon running generate_table8.py, which also generates the table containing the maximum logical wire lengths, as it appeared in the submitted version of the paper.

The obtained maximum lengths should be included in tech.py (Those produced at the time of writing the paper are already embedded in the file).

Please note that, in some cases, the determined buffer sizes may be well matched to the fully modeled loads. To allow appropriate output signal transitions, in these cases, the sizes were manually changed by a small amount. The changes are documented in buf_cache/notes.txt and the entire cache as it appeared when writing the paper is included in the repository.

Architecture Generation

VPR architecture and routing-resource graph descriptions are generated by generate_architecture/arc_gen.py. The script also provides more accurate modeling of global wire delays, including the wires connecting the LUT outputs to the switch-block multiplexers, etc. Many options are provided, whether as command line arguments or global variables at the top of the source file. For the most part, they should be documented in the code itself, but please note that not all of them are guaranteed to work properly, as some were introduced only for preliminary experimentation and dropped either as not promising, or due to lack of time. To reproduce the results of the paper, there should not be any need to change any of the settings, nor to directly run this script.

Running the Main Experiments

Scripts for running the main experiments are contained in explore/.

Obtaining the Minimum Grid Sizes

The minimum FPGA grid sizes are obtained from VPR, by running grid_sizer/size.py.

The obtained results should be included in conf.py (Those produced at the time of writing the paper are already embedded in the file).

Exploring Channel Compositions

To generate and rank the various combinations of wire lengths entering the channel composition, run runner_scripts/run_magic.py, without arguments.

Running All VPR Experiments

To generate all architectures for channel compositions in the order provided in the previous step and run implementation of all circuits on them, run loop_cruncher.py, without arguments, and still in the runner_scripts/ directory.

By default, for each technology node, the process will stop once 3 different compositions that manage to successfully place and route all circuits for all cluster sizes are found. This can be changed by changing the parameter num on line 13 of the script.

VPR switches are included in run_vpr.py, whereas the placement seeds are contained in explore/conf.py, where they may be changed.

Finally, the benchmark circuits are contained in benchmarks/.

Collecting the Results

Once the VPR experiments complete, navigate to explore/processing_scripts/ and run collect.py without arguments. If the number of channel compositions explored in the final experiments is changed from the default 3 (see the previous section), parameter wire_no on line 64 of the script needs to be changed accordingly.

Plotting the Results

For plotting the results, paste the dictionary output to the screen by collect.py into plot_fig15.py as res_dict. The dictionary obtained for the final version of the paper is already embedded in the script.

A Note on Runtime

On an Intel(R) Xeon(R) CPU E5-2680 v3-based server, running at 2.50GHz, with 256 GB of RAM, the main experiments take about 18 h to complete, with the default settings embedded in the scripts. The operating system used at the time of running the experiments was CentOS 8 (centos-release-8.2-2.2004.0.2.el8.x86_64). All this should not have any influence on the results, but may be useful for assessing the time needed for rerunning the experiments or extending them.

Troubleshooting

  • Depending on the version of HSPICE (and possibly other specificities of the system used for running the experiments), the measured delays may differ slightly. This could in turn trigger an exception with a message "Negative time!" in some simulations. Other than changing the appropriate buffer size slightly, increasing the source pulse width from 4 ns to 5 ns has been successfully used to resolve this.

  • The timeout set on line 60 of run_magic.py was determined to be appropriate for the setup described here. If the ".sort" files, listing the channel compositions for each technology, ranked by performance (see the paper for the details) contain very few or no architectures, this may be an indicator that the timeout is too small for the current setup. Increasing it should enable more architectures to be assessed successfully.

Intermediate Files

Final architecture files that came out of the exploration presented in the paper are available in final_arcs.zip. Only the smallest generated grids are provided, but the arc_gen.py script can change the grid size without rerunning the HSPICE simulations. The easiest way to achieve this would be to navigate to generate_architecture/ and call the script with the change_grid_dimensions parameter pointing to the architecture to be resized and other files passed as arguments accordingly. For example, a resizing call could look as follows:

python arc_gen.py --K 6 --N 4 --wire_file ../final_arcs/N4_T4/T4_W15.wire --grid_w 30 --grid_h 30 --density 0.5 --tech 4 --arc_name arc_W30_H30.xml --change_grid_dimensions ../final_arcs/N4_T4/arc_T4_N4_W15_W16_H16.xml --physical_square 1 --import_padding ../final_arcs/N8_T4/arc_T4_N8_W15_W12_H12_padding.log

Note that the padding log, containing the information about the number LEN-1 wires added to the channels should point to the file corresponding to the N=8 architecture, as this is the starting point for channel scaling (more details in the paper).

In case you wish to skip RR-graph compression, set COMPRESS_RR on line 286 of the script to False.

Contact

If you find any bugs please open an issue. For all other questions, including getting access to the development branch, please contact Stefan Nikolic (firstname dot lastname at epfl dot ch).

Acknowledgments

We would like to thank Satwant Singh whose help during the artifact evaluation process for FPGA'21 resulted in eliminating several bugs and finding several solutions that made result reproduction easier.