本文整理匯總了Python中sqlalchemy.DATE屬性的典型用法代碼示例。如果您正苦於以下問題:Python sqlalchemy.DATE屬性的具體用法?Python sqlalchemy.DATE怎麽用?Python sqlalchemy.DATE使用的例子?那麽, 這裏精選的屬性代碼示例或許可以為您提供幫助。您也可以進一步了解該屬性所在類sqlalchemy
的用法示例。
在下文中一共展示了sqlalchemy.DATE屬性的15個代碼示例,這些例子默認根據受歡迎程度排序。您可以為喜歡或者感覺有用的代碼點讚,您的評價將有助於係統推薦出更棒的Python代碼示例。
示例1: downgrade
# 需要導入模塊: import sqlalchemy [as 別名]
# 或者: from sqlalchemy import DATE [as 別名]
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.add_column('services_history', sa.Column('sms_sender', sa.VARCHAR(length=11), autoincrement=False, nullable=True))
op.add_column('services', sa.Column('sms_sender', sa.VARCHAR(length=11), autoincrement=False, nullable=True))
op.create_table('notification_statistics',
sa.Column('id', postgresql.UUID(), autoincrement=False, nullable=False),
sa.Column('service_id', postgresql.UUID(), autoincrement=False, nullable=False),
sa.Column('emails_requested', sa.BIGINT(), autoincrement=False, nullable=False),
sa.Column('emails_delivered', sa.BIGINT(), autoincrement=False, nullable=False),
sa.Column('emails_failed', sa.BIGINT(), autoincrement=False, nullable=False),
sa.Column('sms_requested', sa.BIGINT(), autoincrement=False, nullable=False),
sa.Column('sms_delivered', sa.BIGINT(), autoincrement=False, nullable=False),
sa.Column('sms_failed', sa.BIGINT(), autoincrement=False, nullable=False),
sa.Column('day', sa.DATE(), autoincrement=False, nullable=False),
sa.ForeignKeyConstraint(['service_id'], ['services.id'], name='notification_statistics_service_id_fkey'),
sa.PrimaryKeyConstraint('id', name='notification_statistics_pkey'),
sa.UniqueConstraint('service_id', 'day', name='uix_service_to_day')
)
示例2: downgrade
# 需要導入模塊: import sqlalchemy [as 別名]
# 或者: from sqlalchemy import DATE [as 別名]
def downgrade():
op.create_table('dm_datetime',
sa.Column('bst_date', sa.DATE(), autoincrement=False, nullable=False),
sa.Column('year', sa.INTEGER(), autoincrement=False, nullable=False),
sa.Column('month', sa.INTEGER(), autoincrement=False, nullable=False),
sa.Column('month_name', sa.VARCHAR(), autoincrement=False, nullable=False),
sa.Column('day', sa.INTEGER(), autoincrement=False, nullable=False),
sa.Column('bst_day', sa.INTEGER(), autoincrement=False, nullable=False),
sa.Column('day_of_year', sa.INTEGER(), autoincrement=False, nullable=False),
sa.Column('week_day_name', sa.VARCHAR(), autoincrement=False, nullable=False),
sa.Column('calendar_week', sa.INTEGER(), autoincrement=False, nullable=False),
sa.Column('quartal', sa.VARCHAR(), autoincrement=False, nullable=False),
sa.Column('year_quartal', sa.VARCHAR(), autoincrement=False, nullable=False),
sa.Column('year_month', sa.VARCHAR(), autoincrement=False, nullable=False),
sa.Column('year_calendar_week', sa.VARCHAR(), autoincrement=False, nullable=False),
sa.Column('financial_year', sa.INTEGER(), autoincrement=False, nullable=False),
sa.Column('utc_daytime_start', postgresql.TIMESTAMP(), autoincrement=False, nullable=False),
sa.Column('utc_daytime_end', postgresql.TIMESTAMP(), autoincrement=False, nullable=False),
sa.PrimaryKeyConstraint('bst_date', name='dm_datetime_pkey')
)
op.create_index('ix_dm_datetime_yearmonth', 'dm_datetime', ['year', 'month'], unique=False)
op.create_index('ix_dm_datetime_bst_date', 'dm_datetime', ['bst_date'], unique=False)
示例3: downgrade
# 需要導入模塊: import sqlalchemy [as 別名]
# 或者: from sqlalchemy import DATE [as 別名]
def downgrade():
op.add_column('service_permissions',
sa.Column('updated_at', postgresql.TIMESTAMP(), autoincrement=False, nullable=True))
op.create_table('template_statistics',
sa.Column('id', postgresql.UUID(), autoincrement=False, nullable=False),
sa.Column('service_id', postgresql.UUID(), autoincrement=False, nullable=False),
sa.Column('template_id', postgresql.UUID(), autoincrement=False, nullable=False),
sa.Column('usage_count', sa.BIGINT(), autoincrement=False, nullable=False),
sa.Column('day', sa.DATE(), autoincrement=False, nullable=False),
sa.Column('updated_at', postgresql.TIMESTAMP(), autoincrement=False, nullable=False),
sa.ForeignKeyConstraint(['service_id'], ['services.id'],
name='template_statistics_service_id_fkey'),
sa.ForeignKeyConstraint(['template_id'], ['templates.id'],
name='template_statistics_template_id_fkey'),
sa.PrimaryKeyConstraint('id', name='template_statistics_pkey')
)
示例4: test_reflect_select
# 需要導入模塊: import sqlalchemy [as 別名]
# 或者: from sqlalchemy import DATE [as 別名]
def test_reflect_select(table, table_using_test_dataset):
for table in [table, table_using_test_dataset]:
assert len(table.c) == 18
assert isinstance(table.c.integer, Column)
assert isinstance(table.c.integer.type, types.Integer)
assert isinstance(table.c.timestamp.type, types.TIMESTAMP)
assert isinstance(table.c.string.type, types.String)
assert isinstance(table.c.float.type, types.Float)
assert isinstance(table.c.boolean.type, types.Boolean)
assert isinstance(table.c.date.type, types.DATE)
assert isinstance(table.c.datetime.type, types.DATETIME)
assert isinstance(table.c.time.type, types.TIME)
assert isinstance(table.c.bytes.type, types.BINARY)
assert isinstance(table.c['record.age'].type, types.Integer)
assert isinstance(table.c['record.name'].type, types.String)
assert isinstance(table.c['nested_record.record.age'].type, types.Integer)
assert isinstance(table.c['nested_record.record.name'].type, types.String)
assert isinstance(table.c.array.type, types.ARRAY)
rows = table.select().execute().fetchall()
assert len(rows) == 1000
示例5: downgrade
# 需要導入模塊: import sqlalchemy [as 別名]
# 或者: from sqlalchemy import DATE [as 別名]
def downgrade():
op.create_table('provider_statistics',
sa.Column('id', postgresql.UUID(), autoincrement=False, nullable=False),
sa.Column('day', sa.DATE(), autoincrement=False, nullable=False),
sa.Column('service_id', postgresql.UUID(), autoincrement=False, nullable=False),
sa.Column('unit_count', sa.BIGINT(), autoincrement=False, nullable=False),
sa.Column('provider_id', postgresql.UUID(), autoincrement=False, nullable=False),
sa.ForeignKeyConstraint(['provider_id'], ['provider_details.id'], name='provider_stats_to_provider_fk'),
sa.ForeignKeyConstraint(['service_id'], ['services.id'], name='provider_statistics_service_id_fkey'),
sa.PrimaryKeyConstraint('id', name='provider_statistics_pkey')
)
op.create_index('ix_provider_statistics_service_id', 'provider_statistics', ['service_id'], unique=False)
op.create_index('ix_provider_statistics_provider_id', 'provider_statistics', ['provider_id'], unique=False)
示例6: _get_oracle_cdc_client_origin
# 需要導入模塊: import sqlalchemy [as 別名]
# 或者: from sqlalchemy import DATE [as 別名]
def _get_oracle_cdc_client_origin(connection, database, sdc_builder, pipeline_builder, buffer_locally,
src_table_name=None, batch_size=BATCH_SIZE, **kwargs):
kwargs.setdefault('dictionary_source', 'DICT_FROM_ONLINE_CATALOG')
kwargs.setdefault('logminer_session_window', '${10 * MINUTES}')
kwargs.setdefault('db_time_zone', 'UTC')
kwargs.setdefault('maximum_transaction_length', '${1 * MINUTES}')
kwargs.setdefault('initial_change', 'DATE')
if Version('3.14.0') <= Version(sdc_builder.version) < Version('3.16.0'):
# In versions < 3.16 the user has to define a maximum time to look back for a valid dictionary. From
# 3.16 onward this is not required anymore. By default avoid to set an specific duration and use all
# the redo logs instead.
kwargs.setdefault('duration_of_directory_extraction', -1)
if src_table_name is not None:
if Version(sdc_builder.version) >= Version('3.1.0.0'):
tables = [{'schema': database.username.upper(), 'table': src_table_name, 'excludePattern': ''}]
kwargs.setdefault('tables', tables)
else:
kwargs.setdefault('schema_name', database.username.upper())
kwargs.setdefault('tables', [src_table_name])
elif 'tables' not in kwargs.keys():
raise Exception("Either 'tables' or 'src_table_name' must be passed as argument.")
start = _get_current_oracle_time(connection=connection)
kwargs.setdefault('start_date', start.strftime('%d-%m-%Y %H:%M:%S'))
# The time at the oracle db and the node executing the test may not have the exact same time.
# So wait until this node reaches that time (including the timezone offset),
# otherwise validation will fail because the origin thinks the
# start time is in the future.
_wait_until_time(time=start)
logger.info('Start Date is %s', kwargs['start_date'])
oracle_cdc_client = pipeline_builder.add_stage('Oracle CDC Client')
return oracle_cdc_client.set_attributes(buffer_changes_locally=buffer_locally,
max_batch_size_in_records=batch_size,
**kwargs)
示例7: get_audit_ids_by_date
# 需要導入模塊: import sqlalchemy [as 別名]
# 或者: from sqlalchemy import DATE [as 別名]
def get_audit_ids_by_date(cls, target_type, target_id, date):
db = DBSession()
rs = db.query(cls.audit_id).filter(
cls.target_id == target_id,
cls.target_type == target_type,
cast(cls.created_at, DATE) == date
)
return sorted((r[0] for r in rs), reverse=True)
示例8: get
# 需要導入模塊: import sqlalchemy [as 別名]
# 或者: from sqlalchemy import DATE [as 別名]
def get(self, *args, **kwargs):
git_url = self.get_argument('git_url', default=None, strip=True)
relative_path = self.get_argument('relative_path', default=None, strip=True)
hook_name = self.get_argument('hook_name', default=None, strip=True)
search_val = self.get_argument('search_val', default=None, strip=True)
log_list = []
with DBContext('r') as session:
if git_url and relative_path and hook_name:
offset = datetime.timedelta(days=-3)
re_date = (datetime.datetime.now() + offset).strftime('%Y-%m-%d')
hooks_log = session.query(HooksLog.logs_info).filter(HooksLog.git_url == git_url,
HooksLog.relative_path == relative_path,
HooksLog.hook_name == hook_name).filter(
cast(HooksLog.create_time, DATE) > cast(re_date, DATE)).order_by(-HooksLog.id).first()
if hooks_log:
return self.write(dict(code=0, msg='獲取成功', data=hooks_log[0]))
else:
return self.write(dict(code=-2, msg='未找到最近觸發的任務', data=''))
elif search_val:
hooks_log_info = session.query(HooksLog).filter(
or_(HooksLog.git_url.like('{}%'.format(search_val)),
HooksLog.relative_path.like('{}%'.format(search_val)),
HooksLog.logs_info.like('{}%'.format(search_val)))).order_by(-HooksLog.id).all()
else:
hooks_log_info = session.query(HooksLog).order_by(-HooksLog.id).limit(200).all()
for msg in hooks_log_info:
data_dict = model_to_dict(msg)
data_dict['create_time'] = str(data_dict['create_time'])
log_list.append(data_dict)
return self.write(dict(code=0, msg='獲取成功', data=log_list))
示例9: delete
# 需要導入模塊: import sqlalchemy [as 別名]
# 或者: from sqlalchemy import DATE [as 別名]
def delete(self, *args, **kwargs):
data = json.loads(self.request.body.decode("utf-8"))
day_ago = int(data.get('day_ago', 15))
offset = datetime.timedelta(days=-day_ago)
re_date = (datetime.datetime.now() + offset).strftime('%Y-%m-%d')
with DBContext('w', None, True) as session:
session.query(HooksLog).filter(cast(HooksLog.create_time, DATE) < cast(re_date, DATE)).delete(
synchronize_session=False)
return self.write(dict(code=0, msg='刪除{}天前的數據成功'.format(day_ago)))
示例10: downgrade
# 需要導入模塊: import sqlalchemy [as 別名]
# 或者: from sqlalchemy import DATE [as 別名]
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.drop_column("user", "last_update")
op.add_column(
"poll",
sa.Column(
"current_date",
sa.DATE(),
server_default=sa.text("now()"),
autoincrement=False,
nullable=False,
),
)
# ### end Alembic commands ###
開發者ID:Nukesor,項目名稱:ultimate-poll-bot,代碼行數:16,代碼來源:2020_04_09_f1c0140a53d4_add_last_updated_to_user.py
示例11: upgrade
# 需要導入模塊: import sqlalchemy [as 別名]
# 或者: from sqlalchemy import DATE [as 別名]
def upgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.create_table(
"notification",
sa.Column("id", sa.Integer(), nullable=False),
sa.Column("chat_id", sa.BigInteger(), nullable=False),
sa.Column("select_message_id", sa.BigInteger(), nullable=True),
sa.Column("poll_message_id", sa.BigInteger(), nullable=True),
sa.Column(
"created_at", sa.DateTime(), server_default=sa.text("now()"), nullable=False
),
sa.Column(
"updated_at", sa.DateTime(), server_default=sa.text("now()"), nullable=False
),
sa.Column("poll_id", sa.Integer(), nullable=True),
sa.ForeignKeyConstraint(["poll_id"], ["poll.id"], ondelete="cascade"),
sa.PrimaryKeyConstraint("id"),
sa.UniqueConstraint(
"poll_id", "chat_id", name="one_notification_per_poll_and_chat"
),
)
op.create_index(
op.f("ix_notification_poll_id"), "notification", ["poll_id"], unique=False
)
op.add_column("poll", sa.Column("next_notification", sa.DateTime(), nullable=True))
op.alter_column(
"poll",
"due_date",
existing_type=sa.DATE(),
type_=sa.DateTime(),
existing_nullable=True,
)
# ### end Alembic commands ###
示例12: downgrade
# 需要導入模塊: import sqlalchemy [as 別名]
# 或者: from sqlalchemy import DATE [as 別名]
def downgrade():
# ### commands auto generated by Alembic - please adjust! ###
op.alter_column(
"poll",
"due_date",
existing_type=sa.DateTime(),
type_=sa.DATE(),
existing_nullable=True,
)
op.drop_column("poll", "next_notification")
op.drop_index(op.f("ix_notification_poll_id"), table_name="notification")
op.drop_table("notification")
# ### end Alembic commands ###
示例13: test_reflect_dates
# 需要導入模塊: import sqlalchemy [as 別名]
# 或者: from sqlalchemy import DATE [as 別名]
def test_reflect_dates(self):
metadata = self.metadata
Table(
"date_types",
metadata,
Column("d1", sqltypes.DATE),
Column("d2", oracle.DATE),
Column("d3", TIMESTAMP),
Column("d4", TIMESTAMP(timezone=True)),
Column("d5", oracle.INTERVAL(second_precision=5)),
)
metadata.create_all()
m = MetaData(testing.db)
t1 = Table("date_types", m, autoload=True)
assert isinstance(t1.c.d1.type, oracle.DATE)
assert isinstance(t1.c.d1.type, DateTime)
assert isinstance(t1.c.d2.type, oracle.DATE)
assert isinstance(t1.c.d2.type, DateTime)
assert isinstance(t1.c.d3.type, TIMESTAMP)
assert not t1.c.d3.type.timezone
assert isinstance(t1.c.d4.type, TIMESTAMP)
assert t1.c.d4.type.timezone
assert isinstance(t1.c.d5.type, oracle.INTERVAL)
示例14: test_create_table
# 需要導入模塊: import sqlalchemy [as 別名]
# 或者: from sqlalchemy import DATE [as 別名]
def test_create_table(engine):
meta = MetaData()
table = Table(
'test_pybigquery.test_table_create', meta,
Column('integer_c', sqlalchemy.Integer, doc="column description"),
Column('float_c', sqlalchemy.Float),
Column('decimal_c', sqlalchemy.DECIMAL),
Column('string_c', sqlalchemy.String),
Column('text_c', sqlalchemy.Text),
Column('boolean_c', sqlalchemy.Boolean),
Column('timestamp_c', sqlalchemy.TIMESTAMP),
Column('datetime_c', sqlalchemy.DATETIME),
Column('date_c', sqlalchemy.DATE),
Column('time_c', sqlalchemy.TIME),
Column('binary_c', sqlalchemy.BINARY),
bigquery_description="test table description",
bigquery_friendly_name="test table name"
)
meta.create_all(engine)
meta.drop_all(engine)
# Test creating tables with declarative_base
Base = declarative_base()
class TableTest(Base):
__tablename__ = 'test_pybigquery.test_table_create2'
integer_c = Column(sqlalchemy.Integer, primary_key=True)
float_c = Column(sqlalchemy.Float)
Base.metadata.create_all(engine)
Base.metadata.drop_all(engine)
示例15: test_oracle_cdc_client_stop_pipeline_when_no_archived_logs
# 需要導入模塊: import sqlalchemy [as 別名]
# 或者: from sqlalchemy import DATE [as 別名]
def test_oracle_cdc_client_stop_pipeline_when_no_archived_logs(sdc_builder, sdc_executor, database, buffer_locally, use_pattern):
"""
Test for SDC-8418. Pipeline should stop with RUN ERROR when there is no archived log files.
Runs oracle_cdc_client >> trash
"""
db_engine = database.engine
src_table_name = get_random_string(string.ascii_uppercase, 9)
try:
connection = database.engine.connect()
table = _setup_table(database=database, table_name=src_table_name)
logger.info('Using table pattern: %s', src_table_name)
pipeline_builder = sdc_builder.get_pipeline_builder()
oracle_cdc_client = pipeline_builder.add_stage('Oracle CDC Client')
# Obviously past time so there is no archived redo logs for this.
start_date = '30-09-2017 10:10:10'
tables = [{'schema': database.username.upper(), 'table': src_table_name, 'excludePattern': ''}]
oracle_cdc_client.set_attributes(buffer_changes_locally=buffer_locally,
db_time_zone='UTC',
dictionary_source='DICT_FROM_ONLINE_CATALOG',
initial_change='DATE',
logminer_session_window='${10 * MINUTES}',
max_batch_size_in_records=BATCH_SIZE,
maximum_transaction_length='${1 * MINUTES}',
start_date=start_date,
tables=tables)
trash = pipeline_builder.add_stage('Trash')
_wait_until_time(_get_current_oracle_time(connection=connection))
oracle_cdc_client >> trash
pipeline = pipeline_builder.build('Oracle CDC Client Pipeline').configure_for_environment(database)
pipeline.configuration["shouldRetry"] = False
sdc_executor.add_pipeline(pipeline)
# Pipeline should stop with StageExcception
with pytest.raises(Exception):
sdc_executor.start_pipeline(pipeline)
sdc_executor.stop_pipeline(pipeline)
status = sdc_executor.get_pipeline_status(pipeline).response.json().get('status')
assert 'RUN_ERROR' == status
finally:
if table is not None:
table.drop(db_engine)
logger.info('Table: %s dropped.', src_table_name)