Posts

Sort by:
Post not yet marked as solved
0 Replies
20 Views
Our current status is that after three minutes or more in the background, re-opening the app is a restart. Most of the users are claiming that they were automatically redirected to the home page of our app after a certain period of inactivity in the app. I recently upgraded my Xcode version from 12.4 to 15.3. I did not experience the problem with Xcode 12.4. It is an enterprise application, and the majority of users report restart issues. It occurred at random, and the user device contained only our application, with no other app like entertainment or gaming apps. However, I notice that many other apps are running in the background for an extended period of time (such as 20 minutes or 30 minutes). When I open the app, the same page sometimes appears in the background or the app is refreshed (like, Medium) I am not sure how they do it; I follow Apple. The rules did not do anything after entering the background. Is there anything Apple could do? How can I resolve this issue? Or it is default iOS behaviour. Please provide any documentation related to this. Please help me resolve this issue. Note: iOS device Version 15 to 17 is the latest
Posted
by
Post not yet marked as solved
1 Replies
24 Views
I'm getting an issue even unencrypted video playback also failing with status failed. Error Domain=CoreMediaErrorDomain Code=-12927 "(null)" I unable to find any info on above error code. Is there some way to look this up? Sample master M3U8 is shared below. Note: If I use any variant M3U8 then it is working playing fine.
Posted
by
Post not yet marked as solved
0 Replies
20 Views
I am trying to create an iOS app where I need to immediately know when the iPhone is unlocked. Let's say I want to print a log on the Xcode console whenever the phone is unlocked. From my app, how do I detect if the phone is unlocked? Some code pointers will be highly appreciated. I am a newbie in iOS/APP development. It should work even if my app is not running. Is it even possible to do so?
Posted
by
Post not yet marked as solved
0 Replies
15 Views
I registered as a developer with Apple. Since I work alone, I have an individual developer account. Unfortunately, I am unable to create an iOS distribution certificate from Visual Studio (for Windows). The option is grayed out even though my role in App Store Connect is "Admin". Visual Studio also shows me that my user is an admin - yet I don't seem to have rights to create iOS distribution certificates. I can create iOS development certificates without any problems.
Posted
by
Post not yet marked as solved
0 Replies
16 Views
求帮忙,看看什么问题!!! ![](“https://developer.apple.com/forums/content/attachment/99917220-bcbb-4900-b54a-9c6c8d4ac24c” “title=微信图片_20240506110007.png;宽度=1420;高度=271“)
Posted
by
Post not yet marked as solved
0 Replies
17 Views
My mac m1 arm64 computer has a linking problem in the configure command. However I export this in my bash_profile. export LDFLAGS="-L/opt/homebrew/opt/ncurses/lib" export CPPFLAGS="-I/opt/homebrew/opt/ncurses/include" export PKG_CONFIG_PATH="/opt/homebrew/opt/ncurses/lib/pkgconfig:$PKG_CONFIG_PATH" LDFLAGS+=" -L/opt/homebrew/opt/readline/lib" CPPFLAGS+=" -I/opt/homebrew/opt/readline/include" PKG_CONFIG_PATH+=":/opt/homebrew/opt/readline/lib/pkgconfig" LDFLAGS+=" -L/opt/homebrew/opt/lapack/lib" CPPFLAGS+=" -I/opt/homebrew/opt/lapack/include" PKG_CONFIG_PATH+=":/opt/homebrew/opt/lapack/lib/pkgconfig" configure:20728: checking for library containing beep configure:20764: /opt/homebrew/bin/gcc-12 -o conftest -g -O2 -Wall -I/opt/homebrew/opt/ncurses/include -I/opt/homebrew/opt/readline/include -I/opt/homebrew/opt/lapack/include -L/opt/homebrew/opt/ncurses/lib -L/opt/homebrew/opt/readline/lib -L/opt/homebrew/opt/lapack/lib -W -Wall -L/opt/homebrew/lib/ -L/opt/homebrew/Cellar/readline/8.2.1/lib/ -L/opt/homebrew/Cellar/ncurses/6.4/lib/ -L/opt/local/lib -L/usr/local/lib -I/usr/local/include/ -I/opt/local/include conftest.c >&5 ld: warning: directory not found for option '-L/opt/homebrew/Cellar/readline/8.2.1/lib/' ld: warning: directory not found for option '-L/opt/homebrew/Cellar/ncurses/6.4/lib/' ld: warning: directory not found for option '-L/opt/local/lib' Undefined symbols for architecture arm64: "_beep", referenced from: _main in cc2370AN.o ld: symbol(s) not found for architecture arm64 collect2: error: ld returned 1 exit status
Posted
by
Post not yet marked as solved
0 Replies
14 Views
My device information: iMac, Apple M1, 16GB, macOS 14.4.1 My software information: Xcode Version 15.3 (15E204a) My problem: I created a new project, and I just printed a line of log.When I run the project for the first time and the second time, the log can be printed correctly.But when I modified the code and ran the project for the third time, the console output was still the log of the old code. Who has similar experience to help me?
Posted
by
Post not yet marked as solved
0 Replies
41 Views
I am fairly new to 3D model rendering and do not know where to start. I am trying to, ideally with ARKit & RealityKit or SceneKit, do a scan of an environment. This includes: Applying realistic textures to the model. Being able to save it as a .usdz file (to be able to open it within the App itself) Once it is save do post-processing measurements within the model. I would prefer to accomplish this feature by using a mesh, instead of the pointCloud that is used in the sample project of apple. Would this be doable using Apple's APIs and on a mobile device or would it be necessary to use a third party program? I have managed to create a USDZ file using SceneKit's .scene.write(to:,delegate:) method. However the saved file is a "single object" and it is not possible to use raycasting to do post-processing measurements in the model.
Posted
by
Post not yet marked as solved
0 Replies
14 Views
I have a full app and I have done quite a number of uploads and TestFlights. But after I add app clip to my full app, when I archive, the iOS build will show (processing) but after I refresh the webpage, the build that was shown to be processing previously vanished from the iOS build. This is the image showing the new iOS build: 19 (processing) This is the image showing the iOS builds after I have refreshed the webpage: As you can see the build 19 no longer appears. Now I don't know how to solve this problem. Can anyone help ?
Posted
by
Post not yet marked as solved
0 Replies
9 Views
Hello, I'm applying to the development program. However, an unknown error appears in the account. Can you help identify this error? We are unable to process your request. An unknown error occurred. The error will appear after filling out the form in Enroll Today and selecting as Sole Ent. I wrote to support 5 days ago, but there was no response.
Posted
by
Post not yet marked as solved
0 Replies
11 Views
Hi, just generated a HDR10 MVHEVC file, mediainfo is below: Color range : Limited Color primaries : BT.2020 Transfer characteristics : PQ Matrix coefficients : BT.2020 non-constant Codec configuration box : hvcC+lhvC then generate the segment files with below command: mediafilesegmenter --iso-fragmented -t 4 -f av_1 av_new_1.mov then upload the segment files and prog_index.m3u8 to web server. just find that can not play the HLS stream on Safari... the url is http://ip/vod/prog_index.m3u8 just checked that if i remove the tag Transfer characteristics : PQ when generating the MVHEVC file. above same mediafilesegmenter command and upload the files to web server. the new version of HLS stream is can play on Safari... Is there any way to play HLS PQ video on Safari. thanks.
Posted
by
Post not yet marked as solved
1 Replies
16 Views
Hello, I use Preview to quickly test functionalities. However I found that the unified logging does not output to Preview console. So I had to do both log.debug and print. Is there a way to enable logging to Preview console?
Posted
by
Post not yet marked as solved
0 Replies
7 Views
Hi, just find that I can not play 3D movie stream: MV-HEVC/UHD/Dolby Vision/Dolby Digital/Dolby Atmos(website is https://developer.apple.com/streaming/examples) when I click View 3D example(fMP4), have below issue: my MacOS is 14.4.1 and M2 chip.
Posted
by
Post not yet marked as solved
1 Replies
16 Views
Hi everyone, I wanted to ask if anybody knows what the current status is about the declaration of required reasons APIs. Before May 1, when I uploaded a new build to the App Store Connect and added it to a group with external testers, I got a notification by email like the following: ITMS-91053: Missing API declaration - Your app’s code in the [...] file references one or more APIs that require reasons, including the following API categories: NSPrivacyAccessedAPICategoryDiskSpace. While no action is required at this time, starting May 1, 2024, when you upload a new app or app update, you must include a NSPrivacyAccessedAPITypes array in your app’s privacy manifest to provide approved reasons for these APIs used by your app’s code. In an article published by Apple (https://developer.apple.com/documentation/bundleresources/privacy_manifest_files/describing_use_of_required_reason_api) it is even statet that that after May 1, apps that do not comply are not accepted by the App Store Connect. According to my interpretation, even the upload should be rejected. I am currently in the process to add a privacy manifest and add the declarations. For testing purposes, I wanted to add the declarations step by step and see where I still need to fix anything. My problem is, that the warnings by apple are not beeing sent anymore. I have uploaded a new build after May 1 with no privacy manifest and therefore no API declarations, it was accepted by App Store Connect and even passed the review for an external testers group. Does anybody have information about the following questions? Did Apple shift the deadline? How can I trigger the warning emails again so that I know what to fix and see, when my app is compliant? Thanks in advance!
Posted
by
Post not yet marked as solved
0 Replies
49 Views
It seems like this may have been an issue for a while based on what I've seen, but I have added a toolbar item to a textfield keyboard and it doesn't show. The only way I can get it to show is by opening the keyboard, typing something, closing the keyboard, and then reopening it. Anyone have a workaround for this? It's like Apple purposely wants to make it difficult to close the keyboard. TextField("Something here...", text: $text, axis: .vertical) .multilineTextAlignment(.leading) .toolbar { ToolbarItemGroup(placement: .keyboard, content: { Button("Close") { } }) }
Posted
by
Post not yet marked as solved
0 Replies
6 Views
I have recently started using apple shortcuts to connect to a server to control a remote device. This remote device replies using JSON and depending on the request, it replies with a different status message. I would like to be able to read the reply in the shortcut and change the icon and/or the colour of the shortcut on the home screen depending on the reply. This could be used for locking a door where the reply indicates if the door is locked/unlocked by way of a closed/open padlock or by changing the background colour of the shortcut.
Posted
by
Post not yet marked as solved
0 Replies
8 Views
I'm developing an Apple TV OS app using SwiftUI. I'm encountering an issue with the color of the tab item. I want the color of the tab item to turn green when focused; for instance, when the tab item is selected, its background color should be green. Below is my current code, but it's not working. I've tried several other approaches, but none have worked. struct ContentView: View { @State private var selectedTab = 0 var body: some View { ZStack(alignment: .topLeading){ TabView(selection: $selectedTab) { HomeView() .tabItem { AppColors.gradientColor1 Label("Home", image: "home-dash") } .tag(0) ProductsView() .tabItem { Label("Products", image: "open box-search") } .tag(1) SearchView() .tabItem { Label("Search", image: "search") } .tag(2) } .accentColor(Color.green) .background(.black) Image("amazontabbarlogo") } } }
Posted
by
Post not yet marked as solved
0 Replies
45 Views
If my app utilized the RoomPlan api to create a parametric representation of the room, would it open on iPhones that don’t have lidars? I‘m aware the iPhone models that are equipped with lidar are iPhone 12 Pro & Pro Max, iPhone 13 Pro & Pro Max, iPhone 14 Pro & Pro Max, and iPhone 15 Pro & Pro Max.
Posted
by
Post not yet marked as solved
0 Replies
52 Views
I was following the Swiftui tutorial at section 6 (https://developer.apple.com/tutorials/swiftui/creating-and-combining-views) to use a circle image to create an overlapping effect on the map. Turns out that when using a GeometryReader, the bottom padding was not working at all: VStack{ MapView().frame(height:300) CircleImage() .offset(y: -130) .padding(.bottom, -130) VStack(alignment:.leading){ Text( "Turtle Rock" ) .font( .title ) HStack { Text( "Joshua Tree National Park" ) .font( .subheadline ) Spacer() Text( "California" ) .font( .subheadline ) } }.padding(/*@START_MENU_TOKEN@*/10/*@END_MENU_TOKEN@*/) } This is the code for the CicleImage view: GeometryReader(content: { geometry in let _ = print( geometry.size.width ) AsyncImage( url: URL( string: "https://cms.rationalcdn.com/v3/assets/blteecf9626d9a38b03/bltf5486c52361f2012/6144fafd39dff133fc23de9f/img-ios.png" ) ) .frame(width: geometry.size.width) .clipShape( Circle() ).overlay{ Circle().stroke( .white, lineWidth: 4 ) }.shadow( radius: 7 ) })
Posted
by

TestFlight Public Links

Get Started

Pinned Posts

Categories

See all