Merge pull request #24480 from veblush/tls-python
Bump MACOSX_DEPLOYMENT_TARGET to 10.10 for Python
diff --git a/src/ruby/ext/grpc/extconf.rb b/src/ruby/ext/grpc/extconf.rb
index 76b1468..2d99936 100644
--- a/src/ruby/ext/grpc/extconf.rb
+++ b/src/ruby/ext/grpc/extconf.rb
@@ -22,7 +22,7 @@
grpc_config = ENV['GRPC_CONFIG'] || 'opt'
-ENV['MACOSX_DEPLOYMENT_TARGET'] = '10.7'
+ENV['MACOSX_DEPLOYMENT_TARGET'] = '10.10'
if ENV['AR'].nil? || ENV['AR'].size == 0
ENV['AR'] = RbConfig::CONFIG['AR']
diff --git a/tools/run_tests/python_utils/upload_rbe_results.py b/tools/run_tests/python_utils/upload_rbe_results.py
index 62f57dd..93fec98 100755
--- a/tools/run_tests/python_utils/upload_rbe_results.py
+++ b/tools/run_tests/python_utils/upload_rbe_results.py
@@ -287,7 +287,8 @@
if not args.skip_upload:
# BigQuery sometimes fails with large uploads, so batch 1,000 rows at a time.
- for i in range((len(bq_rows) / 1000) + 1):
- _upload_results_to_bq(bq_rows[i * 1000:(i + 1) * 1000])
+ MAX_ROWS = 1000
+ for i in range(0, len(bq_rows), MAX_ROWS):
+ _upload_results_to_bq(bq_rows[i:i + MAX_ROWS])
else:
print('Skipped upload to bigquery.')
diff --git a/tools/run_tests/run_tests.py b/tools/run_tests/run_tests.py
index bdda15f..2135b5f 100755
--- a/tools/run_tests/run_tests.py
+++ b/tools/run_tests/run_tests.py
@@ -1100,7 +1100,7 @@
out.append(
self.config.job_spec(
['test/core/iomgr/ios/CFStreamTests/build_and_run_tests.sh'],
- timeout_seconds=30 * 60,
+ timeout_seconds=60 * 60,
shortname='ios-test-cfstream-tests',
cpu_cost=1e6,
environ=_FORCE_ENVIRON_FOR_WRAPPERS))