QNNPY is an python-based instrument control toolbox for the Quantum Nanostructure and Nanofabrication Group (QNN) at MIT. The primary goal of this package is to simplify and standardize the measurement of superconducting nanowire single photon detectors (SNSPDs) and other superconducting nanoelectronics. This package is split into two main sections: functions, and instruments.
This reposotory does not contain commands that will perform your measurements for you. Its purpose is to organize, standardize, and improve the existing programming infrastructure.
Within qnnpy\instruments
you will find a collection of scripts containing functions that send SCPI commands to each instrument. Please use the following format/capitalization when creating a new instrument script.
agilent_53131a.py #lower_case
class Agilent53131a: #captalize each word
def read(self): #lower_case
...
def run_sweep(self): #lower_case
...
Changing measurement setups is complicated by the different protocols used by each manufacturer. There are two things to strive for that might alleviate this issue. First, is to try and standarize each definition name. If every multimeter has a command read_voltage
there would be no additional changes to our script beyond changing the instrument name. This of course does not work if the instruments have dissimilar abilities. For example we might bias a device with a voltage source or a current source where commands like set_current
would not apply to both. (Second) In this case it makes sense to code in a conditional statement based on the instrument selection. This is cumbersome but generally is uncommon and will only have to be performed once. See the temperature selection in qnnpy\function\snspd.py
for example.
Within functions there are multiple device classes and one class for all base-level functions.
These base-functions contain standard processes like saving, plotting, or logging. This allows us to standardize the parts of these functions that should never be changed and are applicable to every measurement in the lab (like filepaths or file formats).
Base functions contain standard processes like saving, plotting, or logging. This allows us to standardize the parts of these functions that should never be changed and are applicable to every measurement in the lab (like filepaths or file formats).
For example, qf.plot is basically a stripped down version of matplotlib.pyplot.plot with plot.savefig built in.
import qnnpy.functions.functions as qf
import numpy as np
file_path = r'S:\correct_file_path'
x_data = np.arange(-10,10,.5)
y_data = np.tanh(x_data)
qf.plot(x_data,y_data,linestyle='r-', path=file_path)
A device class contains all of the relevant measurements for that device. For example, a snspd class might contain a count-vs-bias measurement or a IV measurement but you would not expect to find S11phase-vs-temperature. For devices with a large number of potential measurements it makes sense to define each measurement as a subclass. Within each subclass will be the method that runs the measurement/sweep and methods that save and plot the data. The save and plot methods within the subclass will build on the base-functions, adding things like axis labels, titles, additional save parameters; any detail specifict to that subclass.
import qnnpy.functions.functions as qf #base-functions
class Snspd: #Class
def __init__(self, configuration_file):
...
class TriggerSweep(Snspd): #Subclass
def run_sweep(self):
...
def save(self):
data_dict = {'trigger_v':self.trigger_v, 'count_rate':self.counts, 'v_bias':self.v_bias}
qf.save(self.properties, 'trigger_sweep', data_dict)
def plot(self):
qf.plot(self.trigger_v,self.counts,
title=self.sample_name+" "+self.device_type+" "+self.device_name,
xlabel = 'Trigger Voltage (V)',
ylabel = 'Counts (cps)',
path = self.full_path,
close=True)
config = r'Q:\somepath\configuration.yml'
c = snspd.TriggerSweep(config)
data = c.run_sweep() #return data if you want
c.plot()
c.save()
Configuration files should be formatted into four sections: User, Device, Measurements, and Instruments. Below is a quick example showing the four sections. These four sections do not explicitly relate to their own dictionary; each measurement and instrument should have their own dictionary, hence the ellipsis.
Here is an example .yml configuration file with four dictionaries.
User:
name:
Save File: #maybe rename this
sample name: SPG000
device type: snspd
device name: A0
some_measurement: #maybe change to same string as subclass
parameter1:
parameter2:
...
some_instrument:
name:
port:
...
PyYAML 5.1 and later
There have been some issues with permissions from the NAS. Ideally the lab computers will have RO access to the network repository. This appears to have been resolved after NAS reboot.
help(module)
for documentation. ex: help(snspd)
Each computer in the lab is connected to the qnn-nas by maping a network drive to the desired folder.
\\18.25.16.44\qnn-repo
is mapped to Q: for Qnn (or R: if Q: is taken)\\18.25.16.44\Data
is mapped to S: for Synology
- Switched (Source) Attenuator compensation in VNA needs to be added to the script.
- Remove reset on PulseTrace module. Still need to capture settings.
Output log is too long and hard to find information. Change output to just the section specific to the measurement and the instruments.- Logging was broken, consider just using open(), open().write, close() for everything.
- Convert all SNSPD bias to either current or voltage (see linearity check and trigger sweep)
- Save all logging as either *.dat or *.txt and build analysis functions. This should help with memory usage.
- Add positive and negative Isw averaging over multiple sweeps to Isw_calc. Might be good to save Isw as a separate .mat file, or just write to DB.
- Add yaml device parameter option. save parameters to data_dict
Move logs to network locationAdd documentation to qnnpy.functionsAdd temperature controllers. still problems with ICEICE controller already has 1 listener (ICE gui) over serial... so we cannot interface. Currently a script checks if the logging feature within the gui is active and pulls the most recent temperature reading. If logging is off it will notify at begining of measurement and will not save old temp.- Temp vs bias. temp vs counts. (difficult for ice)
- qf.plot should be reformatted to better handel arrays/lists
- yaml NAMESPACE
- device type NAMESPACE
- consider moving subclasses out of snspd class. make them stand alone.
all classes/methods should follow CapWord/lower_case convention.- use NAS opposed to google drive stream to store IoT data
consider saving only the configuration for that measurement. eg just the iv_sweep section gets saved when a iv sweep is run.pip install lakeshore
a lakeshore package for all models, we should look into this morecreate documentation for adding new instruments or codes.- ICE temperature can only be read every 10seconds from log. Need to find a way around this.
- Link AJA Sample database with NAS location
- Link PHIDL gds layouts with NAS location. (still not sure how to do this, see qnngds)
qnnmat (collect all scripts libraries from marco/brenden)new repo qnn\qnnmat- create IoT section for cryogenic integration
A collection of scripts containing device specific measurement functions. functions.py
contains the base-level functions used by most other device specific scripts. Using a device specific script, such as snspd.py
, to measure another device will likely cause more headaches then saved lines of code. The best advice I could give is to either: 1) write a completely new script for your device and use the functions.py
for saving and plotting. or 2) copy a function that is close to what you want from snspd.py
to your own script and edit it there. Otherwise, using the snspd.py
script for another device will not allow the necessary flexibility when testing, things will be plotted in an incorrect way, saving dictionaries will be incorrect, etc.
I have not figured out the best way to orgainize everything. Currently all of the SNSPD measurements are in snspd.py
but there are other files like iv_sweep.py
or tc_measurement.py
that are not exactly device specific.
File containing the Snspd class and measurement subclasses.
If you want to measure a snspd using this script the easiest thing to do is to copy the snspd_test_script.py and snspd_example_ICE (config) to your own directory (either a personal location or to the corresponding NAS S:\ drive location). Once copied, update the configuration path, and the configuration data. Then each measurement can be run from the snspd_test_script.py. Your configuration file should only include the instruments you have on, otherwise the script will try and connect to the instruments. You can delete an instrument from the configuration or comment it out.
In spyder I find the below setup to be the most comfortable. A second panel can be created by right clicking the script tap and selecting split vertically (or, Crtl+{).
Compatable Instruments:
- Counter
- Agilent53131a
- Attenuator
- JDSHA9
- Scope
- LeCroy620Zi
- Meter
- Keithley2700
- Keithley2400
- Keithley2001
- Source
- SIM928
- AWG
- Agilent33250a
- VNA
- KeysightN5224a
- Temperature
- Cryocon350
- Cryocon34
- ICE
- DEWAR
Note: code for instrument setup has been moved to functions.Instruments, and instruments are now accessed through inst.scope, inst.source, etc.
This method belongs to the Snspd class and returns the average count rate for a given number of iterations. This function accepts the counting time, iterations, and trigger voltage.
For every iteration the device is biased, the counts are recorded, the voltage is measured, the results are printed, and the bias is turned off. After the specified number of iterations the average count is returned. If SNSPD voltage is greater than 5mV the counts will be printed with # wire switched
but the value will still be included within the average.
Very primitive Tc measurement for the ICE Oxford. Accepts bias voltage and path.
The path must include filename.txt. Bias the device at some value much lower than Ic. This code will bias device, measure voltage, and turn off voltage, every two seconds. The ICE will measure temp every 10 seconds. Slower sweep is better. Keyboard Interrupt to end. File will be saved at path.
Live-updating plotter. Requires IPython to be enabled for interactive shell. Simpily call plot(x, y) and the plot will add your points live. Once you're done, you can save to a png or jpg by calling save()
Instantiation of the class can also takes optional arguments:
title: str - title of the plot
xlabel: str - label for the x axis
ylabel: str - label for the y axis
legend: bool - whether to show the legend or not
legend_loc: str - location for the legend, default is "best", also can be "upper right", "lower left" etc
max_len: int - maximum allowed length of each line in this plot, default is infinite. if the number of lines in one label exceeds this number, the oldest data points get cut off. if you're running a measurement for a very long time, it's best to set this to a number to prevent overusing memory
plot() also optionally takes in a label: str argument to diffrentiate multiple lines and data points, along with most arguments used in the default matplotlib plot() method
save() can take a name, file path, and file type. if no name is provided, a random name based on the current time will be used instead
Basic usage example:
p = LivePlotter()
for x in range(5):
y = x+2
p.plot(x, y)
the data_saver class is used to save data from both LivePlotter and Data classes into a pre-defined organized file structure based on the run configuration defined in the external yaml file. required arguments for the data_saver function are parameters, which is where the yaml file must be passed, and measurement, which is a string defining what measurement is being done (ie: iv_sweep)
measurement data and information will be organized into {meas_path}/{sample_name}/{device_type}/{device_name}/{measurement}, where:
meas_path - root folder, by default S:\SC\Measurements
sample_name - the 'sample name' key under 'Save File' in the yaml file
device_type - the 'device type' key under 'Save File' in the yaml file
device name - the 'device name' key under 'Save File' in the yaml file
measurement - whatever is passed into the 'measurement' argument when the function is being called
other important arguments to include in the function are:
data - the instance of the data class to be saved
inst - the instance of the instruments class to be saved
plot - instance of the LivePlotter class to be saved
optionally, if instead of 'sample name', the 'Save File' key in the yaml file defines a 'sample name 1' and 'sample name 2', and the data or plot arguments is a list instead of a single instance of the class, then data_saver() will recursively call itself for every sample name included in the yaml file (in this example, 2), and element in the data or plot list.
returns a connection to a database. if kwargs is not provided, then simpily returns a connection to qnndb using Owen's username and password
logs data provided in kwargs to a database table
if connection is not provided or None, then simpily uses qnndb with Owen's username and password
Example:
log_data_to_database("measurement_events", connection=None, user='IR', port=1)
This will add a row to the 'measurement_events' table with the 'user' column as 'IR' and 'port' column as 1
allows you to run sql commands in the format UPDATE table_name SET set_col WHERE conditional
easily from the terminal
if connection is not provided / None, then will connect to qnndb with Owen's username and password
Example:
In [1]: update_table('measurement_ids', 'description=stuff', 'id>1 AND id<=3')
Out: UPDATE measurement_ids SET description='stuff' WHERE id>1 AND id<=3
Primitive class-mocking function which creates a mock instance of a class, replicating any function definitions within the class but replacing the functionality of the original functions with a simple print: was called in
This function isn't related to the measurement-taking and saving in the rest of qnnpy, and is only for code-testing, for example creating mock instances of instruments to test code without actually being connected to the instrument.
This class sets up and stores all compatible instruments:
- Counter
- Agilent53131a
- Attenuator
- JDSHA9
- Scope
- LeCroy620Zi
- Meter
- Keithley2700
- Keithley2400
- Keithley2001
- Source
- SIM928
- AWG
- Agilent33250a
- VNA
- KeysightN5224a
- Temperature
- Cryocon350
- Cryocon34
- ICE
- DEWAR
To use the class, store load_config(yaml_configuration_file) into a variable then pass the variable into Instruments, and store the result (ie: instruments = Instruments(load_config(config_file)) ). Instruments can now be accessed with instruments.scope, instruments.awg, instruments.VNA, etc.
If you want to use multiple of the same type of instrument, for example two source, modify your yaml file to include a Source1 and Source2. When instantiating the Instruments class, these two source will be accessible with instruments.source1, instruments.source2. instruments.source will automatically be set to instruments.source1 as well.
If you do not postfix the instrument type in the yaml file with a number, then Instruments will assume you only intend on using one of that type of instrument and will only load one of that type of instrument (different types of instruments will still load). For example, having a Source and Source1 in a yaml file will only load Source, ignoring Source1 or any other Source. If you intend on using multiple instruments, don't start with any number other than 1 as well.
If an instrument fails to connect, then attempting to get that instrument will result in an attribute error. For example, if source fails to connect, then attempting to call instruments.source will yield "AttributeError: 'Instruments' object has no attribute 'source'"
The data class is used to store and save any collected data
Optionally can be used to automatically save data on a specific interval, for long-term data collection. To set up automatic data saving, set the "autosave" argument when instantiating the Data class to True, for example: d = Data(autosave = True). Note that autosaving only works for csv files. Make sure to still call save() after all measurements have finished to save any final measurements which were taken in between the save_increment. Calling save() by default will save to whatever file location was generated when the data class was created, see data.save() documentation below.
Optionally, can also be set up to automatically log to a database. This requires that the table to log to already exists, and that each inputted data key has a corresponding column in the database table.
Other arguments can also be specified:
autosave : bool, optional
When enabled, periodically empties out Data and auto-saves it to the file location provided. The default is False.
Note: If using autosave, remember to still call save() at the end to store any data in the current save_increment that hasn't been transferred yet!
save_increment : int, optional
How often to autosave whenever store() is called. The default is every 128th time store() is called.
path : str, optional
file path to save to. automatically sets up folders if full path doesn't exist. The default is None.
name : str, optional
file name to save to. if a name is already provided in path, it is overridden by this. The default is None.
file_type : str, optional
file type to save to. The default is 'csv'.
preserve_pos_order : bool, optional
if store(v1=1,v2=2) then store(v2=3, v3=4) is called, by default v1
and v4 will be compressed into the first line, while v2 will appear
on lines 1 and 2. Enabling preserve_pos_order will create empty
columns to fix this ordering. The default is False.
connection : mariadb.connection, optional
If you want to auto-log data to a database, then you can set a connection here.
Just remember to run connection.close() after you're done!
table_name : str, optional
database table name
logtime: bool, optional
logs the time in the data dict as 'time' in addition to other variables whenever store() is called
date.store() - stores data into an internal dictionary in the data class. also makes calls to save() and empty() if autosave is enabled
data.empty() - empties out any data but preserves the key names. for example if you ran d.store(voltage=1, temperature=10), then ran d.empty(), the keys voltage and temperature would still exist, but the values 1 and 10 would be emptied out.
data.save() - saves the current contents of the data class. if autosave is being used, then it's likely that some of the contents of the data class have already been transferred into the file, so it's not guarenteed that this will save all data.
Basic usage example:
d = Data()
for i in range(5):
V = take_voltage()
T = take_temperature()
d.store(voltage=V, temperature=T)
d.save()
d.empty() # clear data
if you want to access data in a data class, you can optionally use one of the following:
voltages: list = d.get('voltage')
voltages: list = d.voltage # beware that this will crash if 'voltage' was never passed in d.store()
Date + database example:
import qnnpy.functions.functions as qf
d = qf.Data(table_name = 'measurements', connection=qf.database_connection()) # see database_connection() for more info on this function
for i in range(5):
V = take_voltage()
T = take_temperature()
d.store(voltage=V, temperature=T) # this requires the database table you're committing to to have columns named 'voltage' and 'temperature'
d.connection.close() # make sure to close the connection after you're done! the Data class does not automatically close the connection.
import qnnpy.functions.functions as qf
+any other necessary imports
config = r'YAML\CONFIG\FILE\LOCATION'
properties = qf.load_config(config)
instruments = qf.Instruments(properties)
+any instrument-specific setup, ie instruments.source.setup_read_volt()
d=qf.Data()
p=qf.LivePlotter()
for i in range(100):
V, I = instruments.source.read_voltage_and_current()
d.store(voltage=V, current=I)
p.plot(V, I)
qf.data_saver(properties, 'iv_sweep', meas_path=r'C:\Users\QNN\Documents\Measurements', data=d, plot=p, inst=instruments)
this makes writing your own measurement scripts very easy, but if you still want pre-existing measurement scripts in snspd.py or ntron.py, then those should still work the same.
Example yaml file, note how sample name 1 and sample name 2 are specified
User:
name: TST
Save File:
sample name 1: SPG755
sample name 2: SPG765
device type: wire
device name: TST
Source:
name: 'Keithley2400'
port: "GPIB0::14"
Temperature:
name: 'Cryocon34'
port: 'GPIB0::12'
channel: 'C'
create seperate Data and LivePlotter classes for each sample:
import qnnpy.functions.functions as qf
sample_1_data = qf.Data()
sample_2_data = qf.Data()
sample_1_plot = qf.LivePlotter()
sample_2_plot = qf.LivePlotter()
now when you save, make sure you select the right yaml file, and include both data classes as an list, and both liveplotters as a list:
import qnnpy.functions.functions as qf
config = r'Q:\qnnpy-beta\yml_configs\example_config.yaml'
properties = qf.load_config(config)
qf.data_saver(properties, "test", r"C:\Users\QNN\Documents\Measurements", data = [sample_1_data, sample_2_data], inst = inst, plot = [sample_1_plot, sample_2_plot])
also note that both the "data" and "plot" arguments in data_saver() are optional