aggregate(client, query[, raise_for_status, ...])
|
send a query to the Aggregate API |
batch(client, queries[, raise_for_status])
|
execute multiple queries in a single request using the Batch API |
check_table_exists(client, table_name)
|
check if a table exists in the instance |
copy_table(client, table_name, copy_table_name)
|
copy a table |
create_column(client, table_name, ...)
|
add or replace a column |
create_database(client, schema)
|
create a database using the specified database schema |
create_job(client[, job_endpoint, query, ...])
|
Create a job for a query that takes longer than 30 seconds to run |
create_table(client, table_name, schema)
|
create a table with the specified table name and schema |
delete_column(client, table_name, column_name)
|
delete a column of a table |
delete_database(client)
|
delete the whole database |
delete_entries(client, query[, use_job])
|
Delete the entries according to the criteria given in the query |
delete_table(client, table_name)
|
delete the specified table |
download_table(client, table_name, output_folder)
|
download a table to a NDJSON file or a gzipped NDJSON file |
estimate(client, query[, raise_for_status, ...])
|
send a query to the Estimate API |
evaluate(client, query[, raise_for_status, ...])
|
send a query to the Evaluate API |
generic_query(client, query[, ...])
|
send a query to the Generic Query API |
get_column_schema(client, table_name, ...)
|
get the schema of the specified column |
get_database_schema(client)
|
get the schema of the database |
get_existing_tables(client)
|
get a list of existing tables in the instance |
get_job_result(client, job_id)
|
Get the result of a job with the specified job id |
get_job_status(client, job_id)
|
Get the status of a job with the specified job id |
get_table_schema(client, table_name)
|
get the schema of the specified table |
get_table_size(client, table_name)
|
return the number of entries of the specified table |
get_version(client)
|
get the aito instance version |
initiate_upload_file(client, table_name)
|
Initial uploading a file to a table |
job_request(client[, job_endpoint, query, ...])
|
make a request to an Aito API endpoint using Job |
match(client, query[, raise_for_status, use_job])
|
send a query to the Match API |
modify(client, query[, raise_for_status, ...])
|
perform atomic modifications on table data using the Modify API |
optimize_table(client, table_name[, use_job])
|
optimize the specified table after uploading the data. |
poll_file_processing_status(client, ...[, ...])
|
Polling the file processing status until the processing finished |
predict(client, query[, raise_for_status, ...])
|
send a query to the Predict API |
query_all_entries(client, table_name[, ...])
|
query all entries of the specified table |
query_entries(client, table_name[, offset, ...])
|
query entries of the specified table |
quick_add_table(client, input_file[, ...])
|
Create a table and upload a file to the table, using the default inferred schema |
quick_predict(client, from_table, ...)
|
generate an example predict query to predict a field |
quick_predict_and_evaluate(client, ...)
|
generate an example predict query to predict a field and the corresponding evaluate query |
recommend(client, query[, raise_for_status, ...])
|
send a query to the Recommend API |
relate(client, query[, raise_for_status, ...])
|
send a query to the Relate API |
rename_table(client, old_name, new_name[, ...])
|
rename a table |
search(client, query[, raise_for_status, ...])
|
send a query to the Search API |
similarity(client, query[, ...])
|
send a query to the Similarity API |
trigger_file_processing(client, table_name, ...)
|
Trigger file processing of uploading a file to a table |
upload_binary_file(client, table_name, ...)
|
upload a binary file object to a table |
upload_binary_file_to_s3(...)
|
Upload a binary file to AWS S3 with using the information from initiate_upload_file() |
upload_entries(client, table_name, entries)
|
populate table entries by batches of batch_size |
upload_file(client, table_name, file_path[, ...])
|
upload a file to the specfied table |