- Department of Mathematics, Bar-Ilan University, Ramat Gan, Israel
Liquidity plays a vital role in the financial markets, affecting a myriad of factors including stock prices, returns, and risk. In the stock market, liquidity is usually measured through the order book, which captures the orders placed by traders to buy and sell stocks at different price points. The introduction of electronic trading systems in recent years made the deeper layers of the order book more accessible to traders and thus of greater interest to researchers. This paper examines the efficacy of leveraging the deeper layers of the order book when forecasting quoted depth—a measure of liquidity—on a per-minute basis. Using Deep Feed Forward Neural Networks, we show that the deeper layers do provide additional information compared to the upper layers alone.
1. Introduction
Numerous studies confirm the vital role of liquidity in the financial markets. As detailed below, liquidity affects the market stability and trading activity. This is accomplished in a myriad of mechanisms. For instance, a lack of liquidity contributes to fluctuations in stock prices and returns. In addition, liquidity is often used by professional traders to manage risk.
In the stock market, trading activity is managed through the limit order book, which represents a collection of buy and sell orders placed by traders at a variety of price points. The best bid and best ask prices represent the current going market prices at which a placed order is expected to execute immediately or within a short amount of time. Since traders place orders at different price points, the best bid and best ask prices often fluctuate throughout the trading day. For instance, an influx of sell orders at market would quickly exhaust the volume available at the best bid price, thereby exposing the next bid layer, which becomes the best bid layer. This lowers the stock price. Thus, the deeper layers of the order book are constantly exposed throughout the trading day.
Liquidity is often measured through the best bid and best ask prices and volumes, which represent the uppermost layers of the limit order book. As a result, historically, much of the research has focused solely on these upper layers, essentially overlooking the deeper layers in the order book. However, the introduction of electronic trading systems in recent years rendered the examination and study of the deeper layers significantly more practical and thus of much greater interest, especially since multiple studies suggest that the deeper layers harbor valuable information about liquidity. Furthermore, apparently no research has been done on forecasting the quoted depth—a widely accepted measure of liquidity comprised of the best bid and best ask volumes—on a per-minute basis, and only a handful of studies examined the information potential of the deeper layers in the order book in relation to this measure.
In this paper, we set out to examine the efficacy of using the deeper, traditionally hidden layers of the order book in forecasting the quoted depth on a per-minute basis. We utilize Deep Feed Forward Neural Networks for our forecasting algorithm. Deep learning has been steadily growing in popularity, especially in academic studies, which have shown successful applications of deep learning algorithms in a variety of settings (see for example, Libman et al., 2019). Despite this, papers researching applications of deep learning methods to the financial markets are often recent and few in number.
Our results indicate the deeper layers of the order book provide some useful information in predicting the quoted depth on a per-minute basis, especially when compared to the uppermost best bid and best ask layers alone.
2. Literature Review
Liquidity plays an important role in the financial markets. As Autore et al. (2011), when a market lacks liquidity, stock prices often fluctuate widely during trading, with returns following suit. Multiple studies suggest that market liquidity impacts trading activity, confirming the essential role of liquidity in the financial markets. For instance, Lee et al. (1993) suggests that specialists use liquidity to manage the risk associated with information asymmetry. Pronk (2006) suggests that earnings announcement affect both the bid-ask spread as well as the quoted depth (both measures of liquidity further explained below). This information is useful to any trader, but particularly to market makers that use it to schedule trades as part of an overall trading strategy.
Many ways to measure liquidity have been offered and researched over the years. Hasbrouck and Seppi (2001) was one of the earlier studies that offered a few benchmarks to measure liquidity. These included the bid-ask spread—the difference between the best bid and best ask prices—as well as the quoted depth, defined as the sum of the volume of shares available for trading at the best bid and best ask prices. While, (Hasbrouck and Seppi, 2001), as well as later studies, offered a few other measures of liquidity, the two aforementioned measures seem to be the most widely studied. Most of the other benchmarks offered (for instance, by Chordia et al., 2000) were different manipulations of the same data, as well.
Despite the accepted role of quoted depth in measuring liquidity, to our knowledge no research has been done on forecasting this measure on a per-minute basis, which is vital for traders. In contrast, plenty of research has been done on forecasting bid-ask spreads, including Groß-KlußMann and Hautsch (2013), Cattivelli and Pirino (2019), and Curato and Lillo (2015). This is even more noteworthy in light of the fact that the research on quoted depth has often yielded conflicting results—for instance, (Hasbrouck and Seppi, 2001) found that liquidity measures exhibit few co-movements while (Chordia et al., 2000) discovered the opposite.
From the aforementioned research on quoted depth, much has focused on the uppermost best bid and best ask layers, with little attention given to the deeper layers of the limit order book. This again is in contrast to extensive research on the information content of the deeper layers and its impact on prices and the bid-ask spread, which has been the topic of extensive debate in the literature. See, for example, (Mäkinen et al., 2019; Nousi et al., 2019).
Since the introduction of electronic trading systems, some financial markets embarked on the practice of exposing the deeper layers of the limit order book, which was impractical in traditional dealer markets. Studies of the impact of these changes on the markets suggest that the deeper layers contain information that impacts trading decisions. For example, Curato and Lillo (2015) suggests that specialists leverage this information in trades, while Bloomfield et al. (2005) found that traders informed about the deeper layers place more limit orders. Madhavan et al. (2005) found that traders submitted fewer orders on the Toronto Stock Exchange once the latter decided to expose the top four layers, again suggesting that traders leverage the information about deeper layers in trading decisions. The article by Cao et al. (2009) shows that the deeper layers of the limit order book contribute to price discovery.
Furthermore, little research attempted to explore the relevance of Deep Neural Networks (DNN) for the limit order book. This is despite the fact that DNN architecture has been successfully used for regression applications in other fields. Examples include He (2014), Yu and Xu (2014), and He (2017), all of whom used DNN to forecast electricity loads. Specifically, He (2014) focused on evaluating Deep Feed Forward Neural Networks for predicting electricity loads.
Deep Feed Forward Neural Networks are a category of learning algorithms that consist of an input layer, output layer, and deeper layers of the neural network, as explained by Goodfellow et al. (2016). The output layer of the neural network is powered by the deeper layers of the neural network, which are built as a chain of non-linear functions, also known as the activation functions . The deepest layer is considered the first—or input—layer, and the final, most superficial layer is the output layer. The first layer is embedded in the second layer, the second layer is embedded in the third layer, and so forth.
Data flows from the input layer and through each layer of the network until it reaches the output layer. This represents a pattern known as “feed-forward,” since the data flows in one direction only. The input layer receives the feature vector . This structure is shown graphically in Figure 1.
Each layer is comprised of an activation function, which uses a set of parameters or weight vectors. These weight vectors are multiplied by the inputs in order to generate the output that is then fed into the next layer. This chain structure enables DNNs to execute complex input transformations when calculating the output. Adjusting the parameters or weight vectors allows the DNN to improve its prediction accuracy.
Finding the optimal transformation requires to find the best weights. This is accomplished by optimizing a cost function, also known as the loss function, using the gradient descent method. This is achieved with back-propagation (see Rumelhart et al., 1986), a dynamic algorithm that uses memory in order to reduce repetitions in the calculation of the chain rule derivatives of the large number of weights in the chain of the non-linear functions. Since the problem is not convex, there is no guarantee to find the absolute minimum, only a local one. However, results from a large body of recent deep learning research show success in a wide range of real world applications, as described by Schmidhuber (2015).
In this paper, we set out to examine the information content of the limit order book as it relates to the quoted depth. Specifically, we use a varying number of order book layers to forecast the change in log quoted depth of six stocks in the Tel Aviv Stock Exchange for the years 2012, 2013, 2017, and 2018, using DNNs for our forecasting algorithm. Our results show that the change in log quoted depth cannot be reliably predicted with the best bid and best ask layers alone, though the addition of one layer significantly improves the prediction accuracy. This suggests that at least some of the deeper layers contribute information compared with the results of using the best bid and best ask layers alone.
3. Methodology
We obtained the full limit order book data from the Tel Aviv Stock Exchange (TASE) for the years 2012, 2013, 2017, and 2018. This data was comprised of text files organized by date. For each trading day, there were two files: one for orders submitted and the other for transactions executed. The file detailing submitted orders also included cancellation requests.
We used these files to reconstruct the trading activity in the TASE. This enabled us to rebuild the limit order book, which was necessary for our research. Thus, we built a program that kept a running tally of all the bid and ask orders at each price level, and executed a transaction whenever there was a match, updating the market price as well as the tally at the relevant price points on both the bid and ask sides accordingly.
We accounted for each order type, as well. For instance, market orders were executed immediately at the current going price, while limit orders were executed whenever a match occurred. When an Iceberg (ICE) order was executed, the next relevant portion of the order was exposed. Special instructions, such as for stop-loss orders, were taken into account, as well. We confirmed that our program's trading activity matched the transactions executed by the TASE, using the files mentioned above. The ICE and Stop-Loss orders were introduced to TASE in 2014 (Exchange, 2014) so only the data from 2017 and 2018 included these order types.
In Table 1 are several summary statistics on the submitted orders as well as executed transactions by our six chosen stocks.
With the limit order book thus full restored, we were able to recapture its contents at every timestamp throughout the trading day. For instance, we could see that at 9:53 a.m., the order book included certain pending orders at different price points. Thus, we could use this information to calculate the total volume of outstanding orders, both by price level (layer) as well as at all price levels combined.
For the purpose of building our model, we used the data from the reconstructed order book layers. First, we created a full representation of the order book layers, sorted by price from lowest to highest. These are shown by where represents the snapshot of the bid side layers and represents the snapshot of the ask side layers. After each order was submitted, our program captured snapshots of the order book on both the bid and ask side at the same time and labeled these snapshots serially. i represents the number of each one such serial snapshots. Thus, i goes from 1…N, where N indicates the number of orders for one stock for a specific date. For each day, we had N snapshots.
Both and consist of the time, price, and volume of the relevant orders submitted. Specifically, can be represented as follows:
where: ti = The time at which snapshot i was captured.
= The worst-bid price e.g., The lowest price at which a buyer agrees to buy in snapshot i.
= The volume available for trading at the worst-bid price in snapshot i.
= The best-bid price, e.g., the highest price at which a buyer agrees to buy in snapshot i.
= The volume available for trading at the best bid price in snapshot i.
Note that in between and there exist kb − 2 (kb excluding the best bid and worst bid layers) active price layers , e.g., prices at which volume is available for trading.
Correspondingly, can be represented as follows:
where: ti = The time at which snapshot i was captured.
= The best-ask price, e.g., the lowest price at which a buyer agrees to sell in snapshot i.
= The volume available for trading at the best ask price in snapshot i.
= The highest price at which a buyer agrees to sell in snapshot i, e.g., the “worst-ask” price.
= The volume available for trading at the “worst-ask” price in snapshot i.
Note that as in the case of the bid snapshots, in between and there exist k−2 active price layers e.g., prices at which volume is available for trading.
For each order book snapshot as i at time ti, we calculated the order book snapshot 1 min ahead, represented by i+:
Note that in certain cases, such as the last minute of the day, i+ did not exist and in these cases we omitted this feature.
We used this data to create the training and test samples for each of the depth configurations, e.g., number of price layers, that we wanted to use in predicting the change in log quoted depth. The number of layers was represented by m and varied from 1 to 9: m ∈ (1, 2…9).
Finally, the log quoted depth LQD was calculated as:
When calculating the log quoted depth, we used the distance of the price of each layer from the best bid and best ask rather than the actual prices. Thus, the depth of a sample price layer ℓ from the best bid is while the depth of a sample price layers from the best ask is .
We included all of this information in our feature vector i for a given number of layers used for the analysis m. Specifically, each sample was represented as follows:
where: m = The number of layers the used for creating the feature vector.
= The ask volume available for trading at the “best-ask”.
= The ask volume available for trading in layer m the “worst-ask” in the feature vector.
= The bid volume available for trading at the “best-bid”.
= The bid volume available for trading in layer m the “worst-bid” in the feature vector.
= The difference between the second “best-bid” price and the “best-bid” price.
= The difference between the price for the m layer “worst-bid” in the feature vector to the “best-bid” price.
= The difference between the second “best-ask” price and the “best-ask” price.
= The difference between the price for the m layer “worst-ask” in the feature vector to the “best-ask” price.
This feature vector was used to predict the change in log quoted depth for a specific snapshot, or ΔLQDi, which was calculated as follows:
We ran the analysis on 10 different configurations of m, e.g., every number of price layers, in order to discern whether the deeper price layers of the limit order book provide any information when forecasting the change in log quoted depth 1 min ahead.
This analysis was done on a yearly basis for six stocks. For each stock and year, the data was split into three sets: train, dev, and test. We used these three sets to cross-validate the model as follows: the train dataset was used to train the model while the dev dataset was used to choose the best parameters for the neural networks. The train and dev dataset consisted of the first 10.5 (10 and a half) months of the year. To create the dev, we randomly sampled 10% of the train data. Then, we used the last 1.5 (one and a half) months of the year for the test dataset, e.g., out-of-sample. The test was used to measure the model's performance in predicting the change in LQD.
We trained the DNN using the Mean Absolute Error (MAE) loss function on the training test and paused the training when the MAE on the dev dataset stabilized, or stopped decreasing. Then, we calculated the MAE as well as the fraction of times in which our model correctly predicted the direction—positive or negative—of ΔLQD on the test dataset.
As stated above, our goal was to determine whether the deeper price layers of the limit order book provide any information when forecasting ΔLQDi, and if so, the ideal number of layers to use when making such a prediction. We chose to work with Feed Forward DNNs, which have shown success in other regression applications, as mentioned previously. However, because our goal was to determine whether the deeper layers contribute any additional information in the prediction, we did not focus on maximizing model performance—for example, by comparing different models or applying pre-training, as done by He (2017) and He (2014).
We used a Feed Forward Deep Neural Network (DNN) for our model with three layers in the neural network configuration. We chose this configuration as it was recommended by He (2014), who compared different DNN configurations for regression on a problem of a similar size. Our algorithm constructed nine different feature vectors corresponding to each number of order book layers used in the analysis and prediction. Thus, we had a feature vector representing zero order book layers (e.g., just the bid and ask), another feature vector representing one order book layers, another feature vector representing two order book layers, etc. Then, our algorithm fed each of these feature vectors into the DNN. This yielded nine DNNs with similar structure except for the first DNN layer, which corresponds to the length of the feature vector.
For the activation function, we chose RELU, a popular function recommended by Ryu et al. (2017), who compared the results of different activation functions for regression purposes. For all of the experiments, we used the same configurations, varying the number of layers only. This ensured that we would be able to attribute any variation in the results to the number layers of the limit order book used in the prediction.
All of our data processing code was written in Python. The DNN was implemented in Tensorflow. We ran our data processing as well as all of the machine-learning algorithms on servers provided by Google Cloud Platform.
In order to estimate the success in predicting ΔLQDi, we used two metrics: MAE and correct direction (CD). These formulas are shown below:
where:
is the predicted value for ΔLQDi,
yi is the actual value of ΔLQDi, and
N is the number of test samples.
4. Results
The full names and tickers of the stocks are detailed in Table 2.
Table 3 shows the average accuracy of our model when predicting the direction of the change in the log quoted depth. The average prediction accuracy is calculated as an average of the accuracy for all six stocks. The table shows the performance by the number of layers used for each year, with 0 representing the uppermost bid-ask layer. For example, our model was able to identify the direction of the change in the log size 49.33% of the time for 2012 when using only the best bid and ask. The accuracy increased to 66.86% when including one more layer and 67.11% when including two layers. The shaded cells indicate that the improvement was significant at the 95% confidence level.
Table 4 shows the p-value of the change in average prediction accuracy with each additional layer, with 0 representing the uppermost bid-ask layer. For example, for 2012, the improvement in prediction accuracy from 0 to 1 layers was significant at nearly 100% confidence, while the improvement from 1 to 2 layers was significant at nearly 99% confidence. The shaded cells highlight values lower than 5%.
As these tables show, including information from the deeper layers in the algorithm does improve the prediction, compared to the upper-most layer alone (e.g., layer 0). However, the additional improvement in prediction accuracy decreases with layer depth. For instance, the inclusion of information from layer 2, when compared to layer 1, yields substantially smaller improvements than the inclusion of layer 1, when compared to layer 0. Furthermore, we see far less statistically significant values as layer depth increases.
Table 5 shows the average MAE by the number of layers used for each year, with 0 representing the uppermost bid-ask layers. Similar to Table 3, the shaded cells indicate that the improvement was significant at the 95% confidence level. Table 6 shows the p-value of the change in average MAE with each additional layer, with the shaded cells highlighting values lower than 5%.
Figure 2 is a graphical depiction of the 2012 averages from Table 3, showing the accuracy by layer and stock. As illustrated in Table 3, our model cannot reliably predict the change in log quoted depth from the uppermost layer (best bid and best ask) alone. However, the addition of just one layer improves the prediction accuracy significantly. Further layers may improve the prediction accuracy further although not always. This suggests that most of the information appears in just one layer below the best bid-ask.
Figure 3 is a graphical representation of the 2012 averages from Table 5. Figures 4, 5 show comparisons across subsets of layers.
5. Conclusion
In this paper, we set out to examine the information content of the deeper layers in the limit order book in the context of predicting the change in log quoted depth on a per-minute basis throughout the trading day. We compared the predictive power of the best bid and best ask combined with additional deeper layers with that of the best bid and best ask layers alone. Our results indicate that the change in log quoted depth cannot be reliable predicted with the best bid and best ask layers alone, and that the addition of the deeper layers substantially improves the prediction accuracy. We therefore conclude that the deeper layers of the order book possess valuable information in the context of liquidity, a finding that is supported by other studies, as well. Understanding the relevance of the deeper layers in the limit order book is especially relevant these days with the introduction of electronic trading markets, which made the open display of the deeper layers more widespread. As the practice of exposing the deeper layers of the order book becomes more prevalent, future research might consider a more detailed comparison of the predictive efficacy of different number of layers. Additionally of interest would be an analysis of multiple trading markets worldwide to examine whether factors such as trading activity or culture affect the predictive power of the deeper layers. For instance, in less busy markets such as TASE, the residual utility of each additional layer might diminish faster than in more active markets such as NYSE.
Data Availability Statement
The data analyzed in this study is subject to the following licenses/restrictions: The dataset involves limit order book trading data from the Tel Aviv Stock Exchange (TASE). Requests to access these datasets should be directed to micky@tase.co.il.
Author Contributions
As a primary researcher, DL was responsible for making the necessary contacts to obtain and collect the relevant data, also wrote the code to process the relevant data and research different algorithms, comparing results, and authoring most of the paper. SH offered substantial guidance throughout the entire project, proposing additional methods, and ways to experiment with the data. During the authoring stage, he initiated numerous thoughtful comments to help refine the paper and ensure its fit with the institution's academic standards. MS supervised the project, and as such she was instrumental in the ideation stage as well as providing overall direction and verifying all mathematical calculations, also facilitated access to crucial resources, including the data sources as well as several colleagues that served as valuable advisors and mentors throughout the project, and contributed essential feedback to the process that proved critical to obtaining quality results in a timely manner. All authors contributed to the article and approved the submitted version.
Funding
The authors declare that this study received funding from Google Cloud in the form of computation resources. The funder was not involved in the study design, collection, analysis, interpretation of data, the writing of this article or the decision to submit it for publication.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
References
Autore, D. M., Billingsley, R. S., and Kovacs, T. (2011). The 2008 short sale ban: Liquidity, dispersion of opinion, and the cross-section of returns of us financial stocks. J. Bank. Fin. 35, 2252–2266. doi: 10.1016/j.jbankfin.2011.01.026
Bloomfield, R., O'Hara, M., and Saar, G. (2005). The “make or take” decision in an electronic market: evidence on the evolution of liquidity. J. Fin. Econ. 75, 165–199. doi: 10.1016/j.jfineco.2004.07.001
Cao, C., Hansch, O., and Wang, X. (2009). The information content of an open limit-order book. J. Futures Mark. 29, 16–41. doi: 10.1002/fut.20334
Cattivelli, L., and Pirino, D. (2019). A SHARP model of bid-ask spread forecasts. Int. J. Forecast. 35, 1211–1225. doi: 10.1016/j.ijforecast.2019.02.008
Chordia, T., Roll, R., and Subrahmanyam, A. (2000). Commonality in liquidity. J. Fin. Econ. 56, 3–28. doi: 10.1016/S0304-405X(99)00057-4
Curato, G., and Lillo, F. (2015). Modeling the coupled return-spread high frequency dynamics of large tick assets. J. Stat. Mech. Theory Exp. 2015:P01028. doi: 10.1088/1742-5468/2015/01/P01028
Exchange, T. T. A. S. (2014). New Order Types Designed to Make Trading More Efficient and Enhance Liquidity. Available online at: https://info.tase.co.il/Eng/NewsandEvents/PRArchive/2014/Pages/PR_20141201_2.aspx
Goodfellow, I., Bengio, Y., and Courville, A. (2016). Deep Learning. Cambridge, MA: MIT Press. Available online at: https://www.deeplearningbook.org/
Groß-Klußmann, A., and Hautsch, N. (2013). Predicting bid-ask spreads using long-memory autoregressive conditional poisson models. J. Forecast. 32, 724–742. doi: 10.1002/for.2267
Hasbrouck, J., and Seppi, D. J. (2001). Common factors in prices, order flows, and liquidity. J. Fin. Econ. 59, 383–411. doi: 10.1016/S0304-405X(00)00091-X
He, W. (2014). Deep neural network based load forecast. Comput. Model. N. Technol. 18, 258–262. Available online at: http://cmnt.lv/upload-files/ns_8art43_CMNT1803-34_N289_He.pdf
He, W. (2017). Load forecasting via deep neural networks. Proc. Comput. Sci. 122, 308–314. doi: 10.1016/j.procs.2017.11.374
Lee, C. M. C., Mucklow, B., and Ready, M. J. (1993). Spreads, depths, and the impact of earnings information: an intraday analysis. Rev. Fin. Stud. 6, 345–374. doi: 10.1093/rfs/6.2.345
Libman, D., Haber, S., and Schaps, M. (2019). Volume prediction with neural networks. Front. Artif. Intell. 2:21. doi: 10.3389/frai.2019.00021
Madhavan, A., Porter, D., and Weaver, D. (2005). Should securities markets be transparent? J. Fin. Mark. 8, 265–287. doi: 10.1016/j.finmar.2005.05.001
Mäkinen, Y., Kanniainen, J., Gabbouj, M., and Iosifidis, A. (2019). Forecasting jump arrivals in stock prices: new attention-based network architecture using limit order book data. Quant. Fin. 19, 2033–2050. doi: 10.1080/14697688.2019.1634277
Nousi, P., Tsantekidis, A., Passalis, N., Ntakaris, A., Kanniainen, J., Tefas, A., et al. (2019). Machine learning for forecasting mid-price movements using limit order book data. IEEE Access 7, 64722–64736. doi: 10.1109/ACCESS.2019.2916793
Pronk, M. (2006). The impact of intraday timing of earnings announcements on the bid-ask spread and depth. J. Account. Audit. Fin. 21, 27–54. doi: 10.1177/0148558X0602100104
Rumelhart, D. E., Hinton, G. E., and Williams, R. J. (1986). Learning representations by back-propagating errors. Nature 323, 533–536. doi: 10.1038/323533a0
Ryu, S., Noh, J., and Kim, H. (2017). Deep neural network based demand side short term load forecasting. Energies 10:3. doi: 10.3390/en10010003
Schmidhuber, J. (2015). Deep learning in neural networks: an overview. Neural Netw. 61(Suppl C), 85–117. doi: 10.1016/j.neunet.2014.09.003
Keywords: limit order book, quoted depth, feed forward, deep learning, deep learning—artificial neural network, deep feedforward, deep feed forward neural network, feed forward algorithm
Citation: Libman D, Haber S and Schaps M (2021) Forecasting Quoted Depth With the Limit Order Book. Front. Artif. Intell. 4:667780. doi: 10.3389/frai.2021.667780
Received: 14 February 2021; Accepted: 06 April 2021;
Published: 11 May 2021.
Edited by:
Dror Y. Kenett, Johns Hopkins University, United StatesReviewed by:
Wei-Xing Zhou, East China University of Science and Technology, ChinaDiana Aldea Mendes, University Institute of Lisbon (ISCTE), Portugal
Copyright © 2021 Libman, Haber and Schaps. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Daniel Libman, daniel.libman@biu.ac.il