Play all audios:
ABSTRACT Data-driven algorithms—such as signal processing and artificial neural networks—are required to process and extract meaningful information from the massive amounts of data currently
being produced in the world. This processing is, however, limited by the traditional von Neumann architecture with its physical separation of processing and memory, which motivates the
development of in-memory computing. Here we report an integrated 32 × 32 vector–matrix multiplier with 1,024 floating-gate field-effect transistors that use monolayer molybdenum disulfide as
the channel material. In our wafer-scale fabrication process, we achieve a high yield and low device-to-device variability, which are prerequisites for practical applications. A statistical
analysis highlights the potential for multilevel and analogue storage with a single programming pulse, allowing our accelerator to be programmed using an efficient open-loop programming
scheme. We also demonstrate reliable, discrete signal processing in a parallel manner. SIMILAR CONTENT BEING VIEWED BY OTHERS DEMONSTRATION OF 4-QUADRANT ANALOG IN-MEMORY MATRIX
MULTIPLICATION IN A SINGLE MODULATION Article Open access 03 October 2024 AN IN-MEMORY COMPUTING ARCHITECTURE BASED ON TWO-DIMENSIONAL SEMICONDUCTORS FOR MULTIPLY-ACCUMULATE OPERATIONS
Article Open access 07 June 2021 A CROSSBAR ARRAY OF MAGNETORESISTIVE MEMORY DEVICES FOR IN-MEMORY COMPUTING Article 12 January 2022 MAIN Over the past decade, billions of sensors from
connected devices have been used to translate physical signals and information to the digital world. Due to their limited computing power, sensors integrated into embedded remote devices
often transmit raw and unprocessed data to their hosts. However, the high energy cost of wireless data transmission1 affects device autonomy and data transmission bandwidth. Improving their
energy efficiency could open a new range of applications and reduce their environmental footprint. Furthermore, data processing will move from remote hosts to local sensor nodes; therefore,
data transmission would be limited to structured and valuable data, which is desirable for such purposes. The von Neumann architecture—in which memory and logic units are separate—is seen as
the critical factor limiting the efficiency of computing systems in general devices and particularly in edge-based devices. The separation between processing and memory imposed by the von
Neumann architecture requires the data to be sent back and forth between the two during data and signal processing or inference in neural networks. This data communication between memory and
processing units already accounts for one-third of the energy spent in scientific computing2. To overcome the von Neumann communication bottleneck3,4, in-memory computing architectures—in
which memory, logic and processing operations are collocated—are being explored. Processing-in-memory devices are especially suitable for performing vector–matrix multiplication, which is a
key operation for data processing and the most intensive calculation in machine-learning algorithms. By taking advantage of the memory’s physical layer to perform the multiply–accumulate
(MAC) operation, this architecture overcomes the von Neumann communication bottleneck. So far, this processing strategy has been used in applications such as solving linear5,6 and
differential equations7, signal and image processing8 and artificial neural network accelerators9,10,11,12. However, the search for the best materials and devices for this type of processor
is still ongoing. Several devices have been studied for in-memory computing, including standard flash memories, emerging resistive random-access memories and ferroelectric
memories3,13,14,15,16,17,18. More recently, two-dimensional (2D) materials have shown promise in the field of beyond-complementary metal–oxide–semiconductor (CMOS) devices19,20,21,22,23,24,
as well as in-memory and in-sensor computing25,26,27,28. Due to their atomic-scale thickness, floating-gate field-effect transistors (FGFETs) based on monolayer molybdenum disulfide (MoS2)
offer high sensitivity to charge variations in the floating gate and reduced cell-to-cell interference. Such devices could be scaled down to sub-100 nm lengths without loss of
performance27,29,30. Moreover, the van der Waals nature of MoS2 allows devices based on these materials to be integrated into the back-end-of-line31. This would allow processors to be
fabricated with multiple levels of memory cores directly integrated with the required interfaces, creating dense in-memory networks. FGFETs based on MoS2 have been used in logic-in-memory32
and in-memory computing as well as as the main building blocks of perceptron layers27,33 where they are projected to offer more than an order of magnitude improvement in power efficiency
compared with CMOS-based circuits30. These demonstrations have highlighted the promise of 2D materials for in-memory computing, but further progress and practical applications require
wafer-scale fabrication and large-scale or very-large-scale system integration. Currently, demonstrations of the wafer-scale and large-scale integration of 2D-semiconducting-materials-based
circuits have been limited to photodetectors34,35,36,37 or traditional analogue and digital integrated circuits38,39,40,41,42; hardware implementations43 with full-wafer and large-scale
system integration involving 2D-materials-based non-volatile memories that can perform computation are missing. In this Article, we report a chip containing a 32 × 32 FGFET matrix with 1,024
memory devices per chip and an 83.1% yield. The working devices show similar _I_DS versus _V_G characteristics and hysteresis. During fabrication, we use wafer-scale metal–organic
chemical-vapour-deposited (MOCVD) monolayer MoS2 as the channel material, and the entire fabrication process is carried out in a 4-inch line cleanroom. We also demonstrate multibit data
storage in each device with a single programming pulse. Finally, we show that our devices can be used in in-memory computing by performing discrete signal processing with different kernels
in a highly parallelized manner. MEMORY MATRIX We approach in-memory computing by exploiting charge-based memories using monolayer MoS2 as a channel material. Specifically, we fabricated
FGFETs to take advantage of the electrostatic sensitivity of 2D semiconductors19. To enable the realization of larger arrays, we organized our FGFETs in a matrix in which we can address
individual memory elements by carefully choosing the corresponding row and column. Figure 1a,b shows a three-dimensional rendering of the memory matrix and the detailed structure of each
FGFET, respectively. The use of a matrix configuration allows a denser topology and directly corresponds to performing vector–matrix multiplications. Our memories are controlled by local 2
nm/40 nm Cr/Pt gates fabricated in a gate-first approach. This allows us to improve the growth of the dielectric by atomic layer deposition38 and to minimize the number of processing steps
that the 2D channel is exposed to, resulting in an improved yield. The floating gate is a 5 nm Pt layer sandwiched between 30 nm HfO2 (block oxide) and 7 nm HfO2 (tunnel oxide). Next, we
etch vias on the HfO2 to electrically connect the bottom metal (M1) and top metal (M2) layers. This is required for routing the source and drain signals without an overlap. Wafer-scale
MOCVD-grown MoS2 is transferred on top of the gate stack and etched to form the transistors’ channels. Supplementary Figs. 1 and 2 provide details about the material quality and
characterization. Finally, 2 nm/60 nm Ti/Au is patterned and evaporated on top, forming the transistors’ drain–source contacts as well as the second metal layer. Methods provides further
details about the fabrication and Supplementary Figs. 3–8 show the characterization details. Figure 1c shows the optical image of the fabricated chip containing 32 rows and 32 columns for a
total of 1,024 memories. In the image, source channels are accessed from the bottom; drain channels, from the right; and gate channels, from the left. Our memories are based on standard
flash memories. The memory mechanism relies on shifting the neutral threshold voltage (_V_TH0) by changing the number of charges in the trapping layer (Δ_Q_), that is, the platinum floating
gate in our case. When a high positive/negative bias is applied to the gate, the band alignment starts favouring the tunnelling in/out of electrons from the semiconductor to the floating
gate, changing the carrier concentration in the trapping layer. We define our memory window (Δ_V_TH) by taking the difference between the threshold voltage from the forward and reverse
paths, which are taken at a constant current level. Our previous work verified the programming mechanism by fitting our experimental curves in a device simulation model27,29. Since the
memory effect entirely relies on a charge-based process, flash memories tend to have better reliability and reproducibility than emerging memories that are material dependent such as
resistive random-access memories and phase-change memories3. We designed and manufactured a custom device interface board to facilitate the characterization of the memory array
(Supplementary Figs. 9 and 10 provide a detailed description). Figure 1d shows the _I_DS versus _V_G sweeps performed for each device. The fabrication presents a yield of 83.1% and the
devices are statistically similar (Supplementary Section 4). The relatively high OFF-state current is due to a lack of resolution of the analogue-to-digital converters used in the setup.
High-resolution single-device measurements confirm the typical OFF-state currents on the order of picoamperes. Figure 1e shows the ON and OFF current distribution over the memory matrix.
Both ON and OFF currents are taken at _V_DS = 100 mV, forming two distinct planes. The ON and OFF current shows a good distribution over the entire matrix. Supplementary Figs. 13 and 14 show
further detailed single-device characterization, confirming the performance of devices as memories with good retention and endurance stabilities. We show that the devices have a
statistically similar memory window Δ_V_TH = 4.30 ± 0.25 V. This value is smaller compared with the one extracted from single-device measurements due to the higher slew rates (5 V s–1)
required for the time-effective characterization of 1,024 devices in the matrix. OPEN-LOOP PROGRAMMING The similarity of the devices motivates us to pursue a statistical study of the
memories’ programming behaviour. In the context of in-memory computing, an open-loop programming analysis is fundamental. Standard write–verify approaches may be too time-consuming when
programming a large flash memory array. A statistical understanding of memory states in an open loop is essential to improve the performance and speed. We perform the experiment such that
each device is independently excited by selecting the corresponding row (_i_) and column (_j_). Analogue switches in the device interface board keep a low-impedance path in the selected row
(_i_)/column (_j_) and high impedance in the remaining rows and columns. This ensures that a potential difference is applied only to the desired device, avoiding unwanted programming. For
the same reason, we divide the device programming and reading into two independent stages. During the programming phase, the corresponding gate line (row) and the corresponding source line
(column) are selected and programming pulses with parameters _T_PULSE and _V_PULSE are applied in the gate. Due to the tunnelling nature of the device, only two terminals are required to
generate the band bending needed for charge injection into the floating gate. After the pulse, the gate voltage is changed to _V_READ, which is low enough to prevent reprogramming the memory
state. In the reading phase, the drain line is also connected, and the conductance value is probed by applying voltage _V_DS to the drain. This two-stage procedure is required because we
are using a three-terminal device; therefore, both gate and drain share the same row, and consequently, the entire row is biased when the gate and drain lines are engaged. If high voltages
in the gate were applied when the drain line is connected, the whole row would be reprogrammed, causing a loss of information in the memories. Figure 2a shows the description of this
two-stage programming procedure. For the subsequent measurements, we used _V_READ = −3 V, _V_DS = 1 V and _T_PULSE = 100 ms. Before each measurement, we reset the memories by applying a
positive 10 V pulse, which puts the devices into a low-conductance state. Due to parasitic resistances in the matrix, a linear compensation in the digital gains is applied (Supplementary
Figs. 17 and 18 provide further details). The compensation method improves the programming reliability of the devices by an order of magnitude. We estimate a programming error of 500 errors
per million for programming one bit and having one error per million for programming the erase state. Figure 2b,c shows the distribution of memory states after different pulse intensities,
namely, _V_PULSE = +10 V, −4 V, −6 V, −8 V and −10 V, in both linear and logarithmic representations. We observe that on a linear scale, the increase in the pulse amplitude is accompanied by
a higher memory state value and a larger spread. On the other hand, by analysing the logarithm of the state value, we can see that the memory has well-defined storage states. This leads us
to conclude that this memory has the potential for multivalued storage without write–verify algorithms, especially when used on a logarithmic scale. Figure 2d shows the spatial distribution
of the states on the entire chip. We observe that the memory states create a constant plane value for the different programming voltages, _V_PULSE. Finally, Fig. 2e shows the empirical
cumulative distribution function (ECDF) of the logarithmic representation. These results support the possibility of multivalued programming, as discussed previously, and indicate that the
memory elements can be used for storing analogue weights for in-memory computing. STATES AND VECTOR–MATRIX MULTIPLICATIONS With the open-loop analysis completed (Fig. 3a), we plot the memory
states (<_w_>) as a function of the programming voltage (_V_PROG). We define four equally distributed states (two-bit resolution) to be programmed as discrete weights in the matrix
for the vector–matrix multiplication (Supplementary Fig. 20). To analyse the effectiveness of the processor for performing vector–matrix operations, we compare (Fig. 3b) the normalized
theoretical (_y_THEORY) value with the normalized experimental (_y_EXP) value obtained on several dot-product operations. The linear regression of the experimental points shows a line with
parameters A = 0.988 ± 0.008 and B = −0.129 ± 0.003 for _y_EXP = A × _y_THEORY + B, whereas the shaded area corresponds to a 95% confidence interval. The ideal processor should converge to A
= 1 and B = 0 with a confidence interval that converges to linear fitting. In our case, the processor has a linear behaviour converging to the ideal case, with a large spread and slight
nonlinearity of the experimental values. We explain this behaviour by the non-ideality of the memories and the quantization error due to the limited resolution of the states. This shift in
parameter B can be explained by the intrinsic transimpedance amplifier offset with memory leakage seen at _y_THEORY = 0, but it does not affect the observed linear trend. We conclude that we
can perform MAC operations with reasonable accuracy. This operation is needed for performing diverse types of algorithms, such as signal processing and inference in artificial neural
networks. SIGNAL PROCESSING Next, we configure this accelerator to perform signal processing to demonstrate a real-world scenario and application. For signal processing, the input signal
(_x_) is convoluted with a kernel (_h_), resulting in the processed signal (_y_). Depending on the nature of the kernel elements, different types of processing can be achieved. Here we limit
ourselves to three different kernels that perform low-pass filtering, high-pass filtering and feedthrough. All the kernels run in parallel within a single processing cycle, demonstrating
the efficiency of this processor targeting data-centric problems by parallelized processing. More kernels could be added in parallel, limited only by the size of the matrix. Figure 4a shows
the convolution operation and the different kernels used for processing the input signal. The strategy to encode negative kernel values into the conductance values of the memories is to
split the kernel (_h_) into a kernel with only the positive values (_h_+) and one with the absolute values of the negative numbers (_h_−) and encode only the positive numbers with a direct
relation with the conductance values (_G_). After the processing is realized, the outputs of the positive (_y_+) and negative (_y_−) kernels are subtracted (_y_+ – _y_−), resulting in the
final signal (_y_). Figure 4b shows the comparison between the original weights and the weights transferred into the memory matrix using the previously described open-loop programming
scheme. To simplify the transfer, we normalize the weight values at each kernel by its maximum value. As a result, we observe a good agreement between the original and experimental values.
Next, to verify the effectiveness of processing, we first construct our input signal (_x_) as a sum of sinusoidal waves with different frequencies. In this way, we can easily probe the
behaviour of the filters at different frequencies without creating an overly complex signal. Since the signal has positive and negative values, the signal amplitude must fall within the
linear region of device operation. Thus, we restrict the signal range from −100 to 100 mV at _V_READ = 0. Figure 4c shows the fast Fourier transform of the simulated processed signals (left)
and experimental signals (right). The grey line in both simulated and measured signals is the fast Fourier transform of each kernel, giving a guideline for the predicted behaviour of each
operation. We highlight that the experimental processing of all three filters matches fairly well with the theoretical values as well as the prototype filter. Altogether, large-scale arrays
of FGFETs based on 2D materials could be used for other applications such as image processing and inference with artificial neural networks. CONCLUSIONS We have reported the large-scale
integration of 2D materials as the semiconducting channel in an in-memory processor. We demonstrated the reliability and reproducibility of our devices both in terms of characterization and
statistical similarity of the programming states in open-loop programming. The processor carries out vector–matrix multiplications and illustrates its functionality by performing discrete
signal processing. Our approach could allow in-memory processors to reap the benefits of 2D materials and bring new functionality to edge devices for the Internet of Things. METHODS
WAFER-SCALE MEMORY FABRICATION The fabrication starts with a p-doped silicon substrate with a 270-nm-thick SiO2 insulating layer. The first metal layer and FGFET gates were fabricated by
photolithography using an MLA150 advanced maskless aligner with a bilayer 0.4-µm-thick LOR 5A/ 1.1-µm-thick AZ 1512 resist. The 2 nm/40 nm Cr/Pt gate metals were evaporated using an
electron-beam evaporator under a high vacuum. After resist removal by dimethyl sulfoxide, deionized water and O2 plasma are used to further clean and activate the surface for HfO2
deposition. The 30-nm-thick HfO2 blocking oxide is deposited by thermal atomic layer deposition using TEMAH and water as precursors with the deposition chamber set at 200 °C. The 5 nm Pt
floating gates were patterned by photolithography and deposited using the same process as described previously. With the same atomic layer deposition system, we deposit the 7-nm-thick HfO2
tunnel oxide layer with the same process mentioned before. Next, vias are exposed using a single-layer 1.5-µm-thick ECI 3007 photoresist and etched by Cl2/BCl3 chemistry reactive ion
etching. After the transfer of MoS2 onto the substrate, patterning it with photolithography using a 2-µm-thick nLOF resist and etching by O2 plasma. Drain–source electrodes are patterned by
photolithography and 2 nm/60 nm Ti/Au is deposited by electron-beam evaporation. To increase the adhesion of contacts and MoS2 onto the substrate, a 200 °C annealing step is performed in a
high vacuum. The devices have a width/length ratio of 49.5 μm/3.1 μm. DEVICE PASSIVATION The fabricated device is first wire-bonded onto a 145-pin pin-grid-array chip carrier. The device is
heated inside an Ar glovebox at 135 °C for 12 h, which removes the adsorbed water from the device surface. After in situ annealing in the glovebox, a lid is glued onto the chip carrier using
a high-vacuum epoxy and cured in an Ar atmosphere. This protects the device from oxygen and water. TRANSFER PROCEDURE The MOCVD-grown material is first spin coated with PMMA A2 at 1,500
r.p.m. for 60 s and baked at 180 °C for 5 min. Next, we attach a 135 °C thermal release tape onto the MoS2 sample and detach it from sapphire in deionized water. After this, we dry the film
and transfer it onto the patterned substrate. Next, we bake the stack at 55 °C for 1 h. We remove the thermal release tape by heating it on the hot plate at 130 °C. Next, we immerse the
sample in an acetone bath for cleaning the tape polymer residues. Finally, we transfer the wafer to an isopropanol bath and dry it in air. MOCVD GROWTH Monolayer MoS2 was grown using the
MOCVD method. Mo(CO)6, Na2MoO4 and diethyl sulfide were used as precursors. NaCl was spin coated as a catalyst. A pre-annealed three-inch _c_-plane sapphire wafer with a small off-cut angle
(<0.2°) was used as a growth substrate (UniversityWafer). The chemical vapour deposition reaction was performed using a home-built furnace system with a four-inch quartz tube reactor and
mass flow controllers connected with Ar, H2, O2 and metal–organic precursors (Mo(CO)6 and diethyl sulfide). For the MoS2 crystal growth, a reactor was heated to 870 °C at ambient pressure
for 20 min. ELECTRICAL MEASUREMENTS The electrical measurements were performed using a custom device interface board connected to a CompactRIO (cRIO-9056) running a real-time LabVIEW 2020
server. We installed the NI-9264 (16-channel analogue output), NI-9205 (32-channel analogue inputs) and NI-9403 (digital input/output) modules. DATA AVAILABILITY The data that support the
findings of this study are available via Zenodo at https://doi.org/10.5281/zenodo.8383470. CHANGE HISTORY * _ 12 DECEMBER 2023 A Correction to this paper has been published:
https://doi.org/10.1038/s41928-023-01113-9 _ REFERENCES * Xu, X. et al. Scaling for edge inference of deep neural networks. _Nat. Electron._ 1, 216–222 (2018). Article Google Scholar *
Kestor, G., Gioiosa, R., Kerbyson, D. J. & Hoisie, A. Quantifying the energy cost of data movement in scientific applications. In _2013 IEEE International Symposium on Workload
Characterization (IISWC)_ 56–65 (IEEE, 2013). * Sebastian, A., Le Gallo, M., Khaddam-Aljameh, R. & Eleftheriou, E. Memory devices and applications for in-memory computing. _Nat.
Nanotechnol._ 15, 529–544 (2020). Article Google Scholar * McKee, S. A. Reflections on the memory wall. In _Proc. 1st Conference on Computing Frontiers—CF’04_ 162 (ACM Press, 2004). * Sun,
Z., Pedretti, G., Bricalli, A. & Ielmini, D. One-step regression and classification with cross-point resistive memory arrays. _Sci. Adv._ 6, eaay2378 (2020). Article Google Scholar *
Sun, Z. et al. Solving matrix equations in one step with cross-point resistive arrays. _Proc. Natl Acad. Sci. USA_ 116, 4123–4128 (2019). Article MathSciNet Google Scholar * Zidan, M. A.
et al. A general memristor-based partial differential equation solver. _Nat. Electron._ 1, 411–420 (2018). Article Google Scholar * Li, C. et al. Analogue signal and image processing with
large memristor crossbars. _Nat. Electron._ 1, 52–59 (2018). Article Google Scholar * Lin, P. et al. Three-dimensional memristor circuits as complex neural networks. _Nat. Electron._ 3,
225–232 (2020). Article Google Scholar * Wang, Z. et al. Reinforcement learning with analogue memristor arrays. _Nat. Electron._ 2, 115–124 (2019). Article Google Scholar * Yao, P. et
al. Fully hardware-implemented memristor convolutional neural network. _Nature_ 577, 641–646 (2020). Article Google Scholar * Wang, Z. et al. Fully memristive neural networks for pattern
classification with unsupervised learning. _Nat. Electron._ 1, 137–145 (2018). Article Google Scholar * Khaddam-Aljameh, R. et al. HERMES-Core—a 1.59-TOPS/mm2 PCM on 14-nm CMOS in-memory
compute core using 300-ps/LSB linearized CCO-based ADCs. _IEEE J. Solid-State Circuits_ 57, 1027–1038 (2022). Article Google Scholar * Jung, S. et al. A crossbar array of magnetoresistive
memory devices for in-memory computing. _Nature_ 601, 211–216 (2022). Article Google Scholar * Berdan, R. et al. Low-power linear computation using nonlinear ferroelectric tunnel junction
memristors. _Nat. Electron._ 3, 259–266 (2020). Article Google Scholar * Ielmini, D. & Wong, H.-S. P. In-memory computing with resistive switching devices. _Nat. Electron._ 1, 333–343
(2018). Article Google Scholar * Bavandpour, M., Sahay, S., Mahmoodi, M. R. & Strukov, D. B. 3D-aCortex: an ultra-compact energy-efficient neurocomputing platform based on commercial
3D-NAND flash memories. _Neuromorph. Comput. Eng._ 1, 014001 (2021). Article Google Scholar * Merrikh-Bayat, F. et al. High-performance mixed-signal neurocomputing with nanoscale
floating-gate memory cell arrays. _IEEE Trans. Neural Netw. Learn. Syst._ 29, 4782–4790 (2018). Article Google Scholar * Radisavljevic, B., Radenovic, A., Brivio, J., Giacometti, V. &
Kis, A. Single-layer MoS2 transistors. _Nat. Nanotechnol._ 6, 147–150 (2011). Article Google Scholar * Ciarrocchi, A. et al. Polarization switching and electrical control of interlayer
excitons in two-dimensional van der Waals heterostructures. _Nat. Photon._ 13, 131–136 (2019). Article Google Scholar * Bertolazzi, S., Krasnozhon, D. & Kis, A. Nonvolatile memory
cells based on MoS2/graphene heterostructures. _ACS Nano_ 7, 3246–3252 (2013). Article Google Scholar * Sangwan, V. K. et al. Gate-tunable memristive phenomena mediated by grain boundaries
in single-layer MoS2. _Nat. Nanotechnol._ 10, 403–406 (2015). Article Google Scholar * Shen, P.-C., Lin, C., Wang, H., Teo, K. H. & Kong, J. Ferroelectric memory field-effect
transistors using CVD monolayer MoS2 as resistive switching channel. _Appl. Phys. Lett._ 116, 033501 (2020). Article Google Scholar * Desai, S. B. et al. MoS2 transistors with 1-nanometer
gate lengths. _Science_ 354, 99–102 (2016). Article Google Scholar * Paliy, M., Strangio, S., Ruiu, P. & Iannaccone, G. Assessment of two-dimensional materials-based technology for
analog neural networks. _IEEE J. Explor. Solid-State Computat._ 7, 141–149 (2021). Google Scholar * Feng, X. et al. Self-selective multi-terminal memtransistor crossbar array for in-memory
computing. _ACS Nano_ 15, 1764–1774 (2021). Article Google Scholar * Migliato Marega, G. et al. Low-power artificial neural network perceptron based on monolayer MoS2. _ACS Nano_ 16,
3684–3694 (2022). Article Google Scholar * Mennel, L. et al. Ultrafast machine vision with 2D material neural network image sensors. _Nature_ 579, 62–66 (2020). Article Google Scholar *
Giusi, G., Marega, G. M., Kis, A. & Iannaccone, G. Impact of interface traps in floating-gate memory based on monolayer MoS. _IEEE Trans. Electron Devices_ 69, 6121–6126 (2022). * Cao,
W., Kang, J., Bertolazzi, S., Kis, A. & Banerjee, K. Can 2D-nanocrystals extend the lifetime of floating-gate transistor based nonvolatile memory? _IEEE Trans. Electron Devices_ 61,
3456–3464 (2014). Article Google Scholar * Hu, V. P.-H. et al. Energy-efficient monolithic 3-D SRAM cell with BEOL MoS2 FETs for SoC scaling. _IEEE Trans. Electron Devices_ 67, 4216–4221
(2020). Article Google Scholar * Migliato Marega, G. et al. Logic-in-memory based on an atomically thin semiconductor. _Nature_ 587, 72–77 (2020). Article Google Scholar * Zhu, K. et al.
Hybrid 2D–CMOS microchips for memristive applications. _Nature_ 618, 57–62 (2023). Article Google Scholar * Hinton, H. et al. A 200 ×256 image sensor heterogeneously integrating a 2D
nanomaterial-based photo-FET array and CMOS time-to-digital converters. In _2022 IEEE International Solid-State Circuits Conference (ISSCC)_ 65, 1–3 (IEEE, 2022). * Dodda, A. et al. Active
pixel sensor matrix based on monolayer MoS2 phototransistor array. _Nat. Mater_. 21, 1379–1387 (2022). * Jang, H. et al. An atomically thin optoelectronic machine vision processor. _Adv.
Mater._ 32, 2002431 (2020). Article Google Scholar * Ma, S. et al. A 619-pixel machine vision enhancement chip based on two-dimensional semiconductors. _Sci. Adv._ 8, eabn9328 (2022).
Article Google Scholar * Yu, L. et al. Design, modeling, and fabrication of chemical vapor deposition grown MoS2 circuits with E-mode FETs for large-area electronics. _Nano Lett._ 16,
6349–6356 (2016). Article Google Scholar * Ma, S. et al. An artificial neural network chip based on two-dimensional semiconductor. _Sci. Bull._ 67, 270–277 (2022). Article Google Scholar
* Wang, X. et al. Analog and logic circuits fabricated on a wafer-scale two-dimensional semiconductor. In _2022 International Symposium on VLSI Technology, Systems and Applications
(VLSI-TSA)_ 1–2 (IEEE, 2022). * Polyushkin, D. K. et al. Analogue two-dimensional semiconductor electronics. _Nat. Electron._ 3, 486–491 (2020). Article Google Scholar * Wachter, S.,
Polyushkin, D. K., Bethge, O. & Mueller, T. A microprocessor based on a two-dimensional semiconductor. _Nat. Commun._ 8, 14948 (2017). Article Google Scholar * Chen, S. et al.
Wafer-scale integration of two-dimensional materials in high-density memristive crossbar arrays for artificial neural networks. _Nat. Electron._ 3, 638–645 (2020). Article Google Scholar
Download references ACKNOWLEDGEMENTS We thank Z. Benes (CMI) for help with the electron-beam lithography and R. Chiesa for assistance with the energy-dispersive X-ray measurements. Device
preparation was carried out in the EPFL Centre of MicroNanotechnology (CMI). We thank B. Bartova and R. Therisod (CIME) for device cross-sectioning and transmission electron microscopy
imaging, which were carried out at the EPFL Interdisciplinary Centre for Electron Microscopy (CIME). We acknowledge support from the European Union’s Horizon 2020 research and innovation
programme under grant agreement nos. 829035 QUEFORMAL (to G.M.M., Z.W. and A.K.), 785219 and 881603 (Graphene Flagship Core 2 and Core 3) to A.K. and 964735 (EXTREME-IR) to H.J and A.K.; the
European Research Council (ERC, grant nos. 682332 and 899775, to H.J., M.T. and A.K.); the CCMX Materials Challenge grant ‘Large area growth of 2D materials for device integration’ (to A.R.
and A.K.); and the Swiss National Science Foundation (grant no. 175822, to G.P. and A.K.). AUTHOR INFORMATION AUTHORS AND AFFILIATIONS * Institute of Electrical and Microengineering, École
Polytechnique Fédérale de Lausanne (EPFL), Lausanne, Switzerland Guilherme Migliato Marega, Hyun Goo Ji, Zhenyu Wang, Gabriele Pasquale, Mukesh Tripathi & Andras Kis * Institute of
Materials Science and Engineering, École Polytechnique Fédérale de Lausanne (EPFL), Lausanne, Switzerland Guilherme Migliato Marega, Hyun Goo Ji, Zhenyu Wang, Gabriele Pasquale, Mukesh
Tripathi & Andras Kis * Institute of Bioengineering, École Polytechnique Fédérale de Lausanne (EPFL), Lausanne, Switzerland Aleksandra Radenovic Authors * Guilherme Migliato Marega View
author publications You can also search for this author inPubMed Google Scholar * Hyun Goo Ji View author publications You can also search for this author inPubMed Google Scholar * Zhenyu
Wang View author publications You can also search for this author inPubMed Google Scholar * Gabriele Pasquale View author publications You can also search for this author inPubMed Google
Scholar * Mukesh Tripathi View author publications You can also search for this author inPubMed Google Scholar * Aleksandra Radenovic View author publications You can also search for this
author inPubMed Google Scholar * Andras Kis View author publications You can also search for this author inPubMed Google Scholar CONTRIBUTIONS A.K. initiated and supervised the project.
G.M.M. fabricated the devices, designed/prepared the measurement setup and performed the device characterization and remaining measurements. H.J. and Z.W. grew the 2D materials and assisted
in materials characterization under the supervision of A.R. M.T. performed the high-resolution transmission electron microscopy for the characterization of devices and materials. G.P.
performed the atomic force microscopy imaging and elemental characterization. A.K. and G.M.M. analysed the data. The manuscript was written by G.M.M. and A.K. with input from all authors.
CORRESPONDING AUTHOR Correspondence to Andras Kis. ETHICS DECLARATIONS COMPETING INTERESTS The authors declare no competing interests. PEER REVIEW PEER REVIEW INFORMATION _Nature
Electronics_ thanks Su-Ting Han, Jing-Kai Huang and the other, anonymous, reviewer(s) for their contribution to the peer review of this work. ADDITIONAL INFORMATION PUBLISHER’S NOTE Springer
Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations. SUPPLEMENTARY INFORMATION SUPPLEMENTARY INFORMATION Supplementary Figs. 1–22
and Sections 1–7. RIGHTS AND PERMISSIONS OPEN ACCESS This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation,
distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and
indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to
the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will
need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/. Reprints and permissions ABOUT THIS ARTICLE
CITE THIS ARTICLE Migliato Marega, G., Ji, H.G., Wang, Z. _et al._ A large-scale integrated vector–matrix multiplication processor based on monolayer molybdenum disulfide memories. _Nat
Electron_ 6, 991–998 (2023). https://doi.org/10.1038/s41928-023-01064-1 Download citation * Received: 23 March 2023 * Accepted: 10 October 2023 * Published: 13 November 2023 * Issue Date:
December 2023 * DOI: https://doi.org/10.1038/s41928-023-01064-1 SHARE THIS ARTICLE Anyone you share the following link with will be able to read this content: Get shareable link Sorry, a
shareable link is not currently available for this article. Copy to clipboard Provided by the Springer Nature SharedIt content-sharing initiative