Search

Alexandros Batsakis

from San Francisco, CA
Age ~46

Alexandros Batsakis Phones & Addresses

  • 880 Corbett Ave APT 3, San Francisco, CA 94131
  • Boston, MA
  • 3333 Charles St, Baltimore, MD 21218 (410) 243-7634

Resumes

Resumes

Alexandros Batsakis Photo 1

Principal Architect

View page
Location:
Berkeley, CA
Industry:
Computer Software
Work:
Splunk
Principal Architect

Teradata Jun 2010 - Jul 2014
Software Engineer and Distributed Systems Architect

Netapp Sep 2008 - Jun 2010
Member of Technical Staff

Johns Hopkins Storage Systems Lab Sep 2002 - May 2008
Research Assistant - Ph.d Candidate

Intracom It Services Jun 1999 - Dec 2001
Software Engineer
Education:
The Johns Hopkins University 2002 - 2008
Ph.D, Computer Science
The Johns Hopkins University 2002 - 2008
M.S.E, Computer Science
Ethnikon kai Kapodistriakon Panepistimion Athinon 1996 - 2001
B.S, Informatics and Telecommunications
Skills:
Distributed Systems
Linux
Storage
Big Data
File Systems
Algorithms
Enterprise Software
Hadoop
Mapreduce
High Performance Computing
Databases
Linux Kernel
Software Engineering
C++
C
Python
Shell Scripting
Perl
Cloud Computing
General Awesomeness
Alexandros Batsakis Photo 2

Big Data Engineer

View page
Location:
San Francisco Bay Area
Industry:
Computer Software
Education:
The Johns Hopkins University 2002 - 2008
Ph.D, Computer Science
The Johns Hopkins University 2002 - 2008
M.S.E, Computer Science
Ethnikon kai Kapodistriakon Panepistimion Athinon 1996 - 2001
B.S, Informatics and Telecommunications
Skills:
File Systems
Storage
Distributed Systems
C++
C
MapReduce
Linux
Hadoop
Databases
General Awesomeness
Big Data
Linux Kernel

Publications

Us Patents

Adaptive Scheduling Of Storage Operations Based On Utilization Of Multiple Client And Server Resources In A Distributed Network Storage System

View page
US Patent:
20100076805, Mar 25, 2010
Filed:
Sep 24, 2008
Appl. No.:
12/237307
Inventors:
Alexandros Batsakis - San Francisco CA, US
Arkady Kanevsky - Swampscott MA, US
James Lentini - Woburn MA, US
Thomas Talpey - Stow MA, US
Assignee:
NetApp, Inc. - Sunnyvale CA
International Classification:
G06Q 30/00
G06F 17/30
G06F 15/16
G06Q 50/00
G06Q 10/00
G06F 9/46
US Classification:
705 8, 718102, 705 26, 707E1701, 707E17032
Abstract:
Scheduling operations such as asynchronous file system operations in a network storage system is accomplished by applying a bid-price online auction methodology, in which bid (willingness-to-pay) values and price (cost) values are dynamically set by storage clients and a storage server, respectively, based on utilization of computing resources. The system provides a framework for adaptively scheduling asynchronous file system operations, managing multiple key resources of the distributed file system, including network bandwidth, server I/O, server CPU, and client and server memory utilization. The system can accelerate, defer, or cancel asynchronous requests to improve application-perceived performance. Congestion pricing via online auctions can be employed to coordinate the use of system resources by clients, so clients can detect shortages and adapt their resource usage.

Cache Aware Searching Based On One Or More Files In Remote Storage

View page
US Patent:
20220292021, Sep 15, 2022
Filed:
Feb 25, 2022
Appl. No.:
17/652635
Inventors:
- San Francisco CA, US
Alexandros Batsakis - San Francisco CA, US
Paul J. Lucas - San Francisco CA, US
Nicholas Robert Romito - San Francisco CA, US
International Classification:
G06F 12/0875
G06F 16/172
G06F 16/951
G06F 16/957
G06F 3/06
G06F 12/0802
G06F 16/14
G06F 12/0862
G06F 12/0866
G06F 12/0868
G06F 12/0871
G06F 12/0873
Abstract:
Embodiments are disclosed for performing cache aware searching. In response to a search query, a first bucket and a second bucket in remote storage for processing the search query. A determination is made that a first file in the first bucket is present in a cache when the search query is received. In response to the search query, a search is performed using the first file based on the determination that the first file is present in the cache when the search query is received, and the search is performed using a second file from the second bucket once the second file is stored in the cache.

Bucket Merging For A Data Intake And Query System Using Size Thresholds

View page
US Patent:
20220261385, Aug 18, 2022
Filed:
Apr 29, 2022
Appl. No.:
17/661510
Inventors:
- San Francisco CA, US
Tianyi Gou - San Francisco CA, US
Alexandros Batsakis - San Francisco CA, US
Abhinav Prasad Nekkanti - Alameda CA, US
Sai Krishna Sajja - Union City CA, US
Jiahan Wang - San Mateo CA, US
International Classification:
G06F 16/22
Abstract:
Systems and methods are disclosed for scalable bucket merging in a data intake and query system. Various components of a bucket manager can be used to monitor recently-created buckets of data in common storage that are associated with a particular tenant and a particular index, apply a comprehensive bucket merge policy to determine groups of buckets that qualify for merging, merge those group of buckets into merged buckets to be stored in the common storage, and update any information associated with the merged buckets and pre-merged buckets. These components may be shared across multiple tenants, and some of these components may be dynamically scalable based on need. This approach may also provide many additional benefits, including improved search performance from merged buckets, efficient resource utilization associated with discriminate merging, and redundancy in case of component failure.

Management Of Distributed Computing Framework Components In A Data Fabric Service System

View page
US Patent:
20210117425, Apr 22, 2021
Filed:
Oct 18, 2019
Appl. No.:
16/657899
Inventors:
- San Francisco CA, US
Jindrich Dinga - Los Altos CA, US
Kieran Cairney - Winnipeg, CA
Manuel Martinez - San Jose CA, US
Arindam Bhattacharjee - Fremont CA, US
Sourav Pal - Foster City CA, US
Alexandros Batsakis - San Francisco CA, US
International Classification:
G06F 16/2453
G06F 16/2458
H04L 9/08
G06F 8/61
H04L 29/08
H04L 12/24
Abstract:
Systems and methods are described for establishing and managing components of a distributed computing framework implemented in a data intake and query system. The distributed computing framework may include a master and a plurality of worker nodes. The master may selectively operate on a search head captain that is chosen from the search heads of the data intake and query system. The search head captain may distribute configuration information for the master and the distributed computing framework to the other search heads, which in turn, may distribute that configuration information to indexers of the data intake and query system. Worker nodes may be selectively activated for operation on the indexers based on the configuration information, and the worker nodes may additionally use the configuration information to contact the master and join the distributed computing framework. This approach may provide numerous benefits, including improved security, flexibility in the selection of worker nodes, and redundancy for failures of physical components of the data intake and query system.

Cache Aware Searching Based On One Or More Files In One Or More Buckets In Remote Storage

View page
US Patent:
20200356487, Nov 12, 2020
Filed:
May 29, 2020
Appl. No.:
16/888320
Inventors:
- San Francisco CA, US
Alexandros Batsakis - San Francisco CA, US
Paul J. Lucas - San Francisco CA, US
Nicholas Robert Romito - San Francisco CA, US
International Classification:
G06F 12/0875
G06F 16/172
G06F 16/951
G06F 16/957
G06F 3/06
G06F 12/0802
G06F 16/14
G06F 12/0862
G06F 12/0866
G06F 12/0868
G06F 12/0871
G06F 12/0873
Abstract:
Embodiments are disclosed for performing cache aware searching. In response to a search query, a first bucket and a second bucket in remote storage for processing the search query. A determination is made that a first file in the first bucket is present in a cache when the search query is received. In response to the search query, a search is performed using the first file based on the determination that the first file is present in the cache when the search query is received, and the search is performed using a second file from the second bucket once the second file is stored in the cache.

Pre-Fetching Files From Buckets In Remote Storage For A Cache Based On File Usage History

View page
US Patent:
20180336215, Nov 22, 2018
Filed:
Jul 30, 2018
Appl. No.:
16/049357
Inventors:
- San Francisco CA, US
Alexandros Batsakis - San Francisco CA, US
Paul J. Lucas - San Francisco CA, US
Nicholas Robert Romito - San Francisco CA, US
International Classification:
G06F 17/30
G06F 12/0875
G06F 3/06
G06F 12/0802
G06F 12/0862
G06F 12/0866
G06F 12/0873
G06F 12/0871
G06F 12/0868
Abstract:
Embodiments are disclosed for a prefetching method that may include copying, in response to a search query, a first bucket from a remote storage to a cache. The first bucket may include first data associated with the search query. The method may further include identifying a first file type associated with a first file in the first bucket. The first file may be associated with a usage status. The method may further include accessing, based on the search query, a second bucket from the remote storage. The second bucket may include second data associated with the search query. The method may further include identifying a second file in the second bucket having the first file type, and copying, in response to the usage status indicating that the first file was used in processing the search query, the second file from the remote storage to the cache.

Cache Aware Searching Based On One Or More Files In One Or More Buckets In Remote Storage

View page
US Patent:
20180336216, Nov 22, 2018
Filed:
Jul 30, 2018
Appl. No.:
16/049609
Inventors:
- San Francisco CA, US
Alexandros Batsakis - San Francisco CA, US
Paul J. Lucas - San Francisco CA, US
Nicholas Robert Romito - San Francisco CA, US
Assignee:
Splunk, Inc. - San Francisco CA
International Classification:
G06F 17/30
G06F 12/0875
G06F 3/06
G06F 12/0802
G06F 12/0862
G06F 12/0866
G06F 12/0873
G06F 12/0871
G06F 12/0868
Abstract:
Embodiments are disclosed for performing cache aware searching. In response to a search query, a first bucket and a second bucket in remote storage for processing the search query. A determination is made that a first file in the first bucket is present in a cache when the search query is received. In response to the search query, a search is performed using the first file based on the determination that the first file is present in the cache when the search query is received, and the search is performed using a second file from the second bucket once the second file is stored in the cache.

Pre-Fetching Data From Buckets In Remote Storage For A Cache

View page
US Patent:
20180196753, Jul 12, 2018
Filed:
Jan 9, 2017
Appl. No.:
15/402105
Inventors:
- San Francisco CA, US
Alexandros Batsakis - San Francisco CA, US
Paul J. Lucas - San Francisco CA, US
Nicholas Robert Romito - San Francisco CA, US
International Classification:
G06F 12/0875
G06F 17/30
Abstract:
Embodiments are disclosed for a prefetching method that may include copying, in response to a search query, a first bucket from a remote storage to a cache. The first bucket may include first data associated with the search query. The method may further include identifying a first file type associated with a first file in the first bucket. The first file may be associated with a usage status. The method may further include accessing, based on the search query, a second bucket from the remote storage. The second bucket may include second data associated with the search query. The method may further include identifying a second file in the second bucket having the first file type, and copying, in response to the usage status indicating that the first file was used in processing the search query, the second file from the remote storage to the cache.
Alexandros Batsakis from San Francisco, CA, age ~46 Get Report